{ "cells": [ { "cell_type": "code", "execution_count": null, "id": "4a567878", "metadata": {}, "outputs": [], "source": [ "import requests\n", "class DownloadEnamine:\n", " \"\"\"\n", " This class is set up to download Enamine REAL database on a remote machine.\n", " Instatiation requires plain ``username`` and ``password``.\n", "\n", " .. code-block::python\n", " de = DownloadEnamine('foo.bar@baz.ac.uk', 'Foo123')\n", " de.download_all('REAL')\n", "\n", " Note, this is copied off the route of the web page and not the Enamine Store API.\n", " Plus the official documentation (emailed Word document) is for the old Store and\n", " no longer applies anyway (plain text username and password in GET header \"Authorization\").\n", "\n", " The URLs pointing to the download pages were copied off manually.\n", " \"\"\"\n", " REAL=[\n", " '2024.07_Enamine_REAL_HAC_25_1B_CXSMILES.cxsmiles.bz2',\n", " ]\n", " LOGIN_URL = 'https://enamine.net/compound-collections/real-compounds/real-database'\n", "\n", " def __init__(self, username, password):\n", " self.sesh = requests.Session()\n", " login_payload = {\n", " 'username': username,\n", " 'password': password,\n", " 'Submit': 'Login',\n", " 'remember': 'yes',\n", " 'option': 'com_users',\n", " 'task': 'user.login'\n", " }\n", " self.sesh.headers.update({'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36'})\n", " response = self.sesh.post(self.LOGIN_URL, data=login_payload)\n", " response.raise_for_status()\n", "\n", " print(\"Login appears successful.\")\n", "\n", " def download_all(self, catalogue='REAL'):\n", " \"\"\"\n", " The URLs of the databases files are in the class attribute of that same catalogue name (i.e. ``REAL``).\n", " \"\"\"\n", " for filename in getattr(self, catalogue):\n", " self.download('REAL', filename)\n", "\n", " def check(self, catalogue='REAL'):\n", " for filename in getattr(self, catalogue):\n", " with self.sesh.get(f'https://ftp.enamine.net/download/{catalogue}/{filename}', stream=True) as r:\n", " r.raise_for_status() # requests.exceptions.HTTPError\n", " for chunk in r.iter_content(chunk_size=8192):\n", " break\n", "\n", " def download(self, catalogue, filename):\n", " \"\"\"\n", " Downloads the ``filename`` of the given ``catalogue``\n", " \"\"\"\n", " with self.sesh.get(f'https://ftp.enamine.net/download/{catalogue}/{filename}', stream=True) as r:\n", " r.raise_for_status()\n", " with open(filename, 'wb') as f:\n", " for chunk in r.iter_content(chunk_size=8192):\n", " f.write(chunk)\n", "\n", "\n", "real_download = DownloadEnamine('taitdang@stanford.edu', 'Z!6CJd2BjQs!y4x')\n", "real_download.download_all('REAL')" ] }, { "cell_type": "code", "execution_count": null, "id": "c2cd8a3d", "metadata": {}, "outputs": [], "source": [ "!source /lfs/skampere1/0/sttruong/miniconda3/etc/profile.d/conda.sh" ] }, { "cell_type": "code", "execution_count": null, "id": "eceda207", "metadata": {}, "outputs": [], "source": [ "export LD_LIBRARY_PATH=\"/lfs/skampere1/0/sttruong/cheapvs_llm/Vina-GPU-2.1/AutoDock-Vina-GPU-2.1/boost_1_77_0/stage/lib:${LD_LIBRARY_PATH}\"" ] }, { "cell_type": "code", "execution_count": null, "id": "86fc4291", "metadata": {}, "outputs": [], "source": [ "conda install jupyter ipykernel\n", "python -m ipykernel install --user --name cheapvs_llm --display-name \"Python (cheapvs_llm)\"\n" ] }, { "cell_type": "code", "execution_count": 1, "id": "56a50433", "metadata": {}, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "Reading SMILES from: smiles_sampled_20k.txt\n", "Writing CSV with column 'SMILES' to: lig_llm.csv\n", "------------------------------\n", "------------------------------\n", "Processing finished.\n", "Successfully wrote: 20354 SMILES strings to CSV.\n", "Output CSV written to: lig_llm.csv\n" ] } ], "source": [ "import csv\n", "import os\n", "import sys\n", "\n", "# --- Configuration ---\n", "input_smiles_file = \"smiles_sampled_20k.txt\" # <-- Your input SMILES text file\n", "output_csv_file = \"lig_llm.csv\" # <-- Your desired output CSV filename\n", "csv_column_name = \"SMILES\" # <-- The header for the column in the CSV\n", "# --- End Configuration ---\n", "\n", "# --- Script Start ---\n", "\n", "# Check if input file exists\n", "if not os.path.isfile(input_smiles_file):\n", " print(f\"ERROR: Input file not found: {input_smiles_file}\")\n", " sys.exit(1)\n", "\n", "print(f\"Reading SMILES from: {input_smiles_file}\")\n", "print(f\"Writing CSV with column '{csv_column_name}' to: {output_csv_file}\")\n", "print(\"-\" * 30)\n", "\n", "count_success = 0\n", "count_skipped = 0\n", "\n", "try:\n", " # Open the input text file for reading ('r')\n", " # Open the output CSV file for writing ('w')\n", " # newline='' is important to prevent blank rows being inserted by csv.writer\n", " with open(input_smiles_file, 'r') as infile, \\\n", " open(output_csv_file, 'w', newline='') as outfile:\n", "\n", " # Create a CSV writer object\n", " csv_writer = csv.writer(outfile)\n", "\n", " # Write the header row\n", " csv_writer.writerow([csv_column_name])\n", "\n", " # Process each line in the input file\n", " for i, line in enumerate(infile):\n", " smiles_string = line.strip() # Remove leading/trailing whitespace (including newline)\n", "\n", " if not smiles_string: # Skip empty lines\n", " print(f\"Warning: Skipped empty line at line number {i+1}\")\n", " count_skipped += 1\n", " continue\n", "\n", " # Write the SMILES string as a row in the CSV\n", " # writerow expects an iterable (like a list), even for a single column\n", " csv_writer.writerow([smiles_string])\n", " count_success += 1\n", "\n", "except IOError as e:\n", " print(f\"ERROR: Could not open or write file. Details: {e}\")\n", " sys.exit(1)\n", "except Exception as e:\n", " # Catch any other unexpected errors during processing\n", " print(f\"An unexpected error occurred processing line {i+1}: {e}\")\n", " sys.exit(1)\n", "\n", "\n", "print(\"-\" * 30)\n", "print(f\"Processing finished.\")\n", "print(f\"Successfully wrote: {count_success} SMILES strings to CSV.\")\n", "if count_skipped > 0:\n", " print(f\"Skipped empty lines: {count_skipped}\")\n", "print(f\"Output CSV written to: {output_csv_file}\")" ] }, { "cell_type": "code", "execution_count": 10, "id": "cdd74781", "metadata": {}, "outputs": [], "source": [ "from admet_ai import ADMETModel\n", "import numpy as np\n", "import pandas as pd\n", "from rdkit import Chem\n", "from rdkit.Chem import (\n", " QED,\n", " Crippen,\n", " Descriptors,\n", " Lipinski,\n", " MolSurf,\n", " RDConfig,\n", " AllChem,\n", " rdMolDescriptors as rdmd,\n", ")\n", "from rdkit.Chem.rdMolDescriptors import CalcFractionCSP3\n", "from rdkit.DataStructs import TanimotoSimilarity\n", "from scipy.stats import norm\n", "from pandarallel import pandarallel\n", "sys.path.append(os.path.join(RDConfig.RDContribDir, \"SA_Score\"))\n", "import sascorer\n", "\n", "def add_property_columns(df_in):\n", " pandarallel.initialize(progress_bar=False, nb_workers=12)\n", " model = ADMETModel(include_physchem=False, num_workers=12)\n", " \"\"\"Add property columns to the DataFrame using pandarallel.\"\"\"\n", " # Convert SMILES to RDKit Mol objects\n", " df_in[\"Mol\"] = df_in[\"SMILES\"].parallel_apply(lambda x: Chem.MolFromSmiles(x))\n", " AliphaticRings = Chem.MolFromSmarts(\"[$([A;R][!a])]\")\n", " df_in[\"MW\"] = df_in[\"Mol\"].parallel_apply(Descriptors.MolWt)\n", " df_in[\"LogP\"] = df_in[\"Mol\"].parallel_apply(Crippen.MolLogP)\n", " df_in[\"numHDonors\"] = df_in[\"Mol\"].parallel_apply(Lipinski.NumHDonors)\n", " df_in[\"numHAcceptors\"] = df_in[\"Mol\"].parallel_apply(Lipinski.NumHAcceptors)\n", " df_in[\"TPSA\"] = df_in[\"Mol\"].parallel_apply(MolSurf.TPSA)\n", " df_in[\"rotBonds\"] = df_in[\"Mol\"].parallel_apply(\n", " lambda mol: rdmd.CalcNumRotatableBonds(\n", " mol, rdmd.NumRotatableBondsOptions.Strict\n", " )\n", " )\n", " df_in[\"Arom\"] = df_in[\"Mol\"].parallel_apply(\n", " lambda mol: len(\n", " Chem.GetSSSR(Chem.DeleteSubstructs(Chem.Mol(mol), AliphaticRings))\n", " )\n", " )\n", " df_in[\"QED\"] = df_in[\"Mol\"].parallel_apply(QED.qed)\n", " df_in[\"SA\"] = df_in[\"Mol\"].parallel_apply(sascorer.calculateScore)\n", " df_in[\"FractionCSP3\"] = df_in[\"Mol\"].parallel_apply(\n", " lambda mol: CalcFractionCSP3(mol)\n", " )\n", "\n", " admet_predictions = model.predict(smiles=df_in[\"SMILES\"].values)\n", " properties = ['Lipinski', 'stereo_centers', 'AMES', 'BBB_Martins', 'Bioavailability_Ma', 'CYP1A2_Veith', 'CYP2C19_Veith', 'CYP2C9_Substrate_CarbonMangels', 'CYP2C9_Veith', 'CYP2D6_Substrate_CarbonMangels', 'CYP2D6_Veith', 'CYP3A4_Substrate_CarbonMangels', 'CYP3A4_Veith', 'Carcinogens_Lagunin', 'ClinTox', 'DILI', 'HIA_Hou', 'NR-AR-LBD', 'NR-AR', 'NR-AhR', 'NR-Aromatase', 'NR-ER-LBD', 'NR-ER', 'NR-PPAR-gamma', 'PAMPA_NCATS', 'Pgp_Broccatelli', 'SR-ARE', 'SR-ATAD5', 'SR-HSE', 'SR-MMP', 'SR-p53', 'Skin_Reaction', 'hERG', 'Caco2_Wang', 'Clearance_Hepatocyte_AZ', 'Clearance_Microsome_AZ', 'Half_Life_Obach', 'HydrationFreeEnergy_FreeSolv', 'LD50_Zhu', 'Lipophilicity_AstraZeneca', 'PPBR_AZ', 'Solubility_AqSolDB', 'VDss_Lombardo', 'Lipinski_drugbank_approved_percentile', 'stereo_centers_drugbank_approved_percentile', 'tpsa_drugbank_approved_percentile', 'AMES_drugbank_approved_percentile', 'BBB_Martins_drugbank_approved_percentile', 'Bioavailability_Ma_drugbank_approved_percentile', 'CYP1A2_Veith_drugbank_approved_percentile', 'CYP2C19_Veith_drugbank_approved_percentile', 'CYP2C9_Substrate_CarbonMangels_drugbank_approved_percentile', 'CYP2C9_Veith_drugbank_approved_percentile', 'CYP2D6_Substrate_CarbonMangels_drugbank_approved_percentile', 'CYP2D6_Veith_drugbank_approved_percentile', 'CYP3A4_Substrate_CarbonMangels_drugbank_approved_percentile', 'CYP3A4_Veith_drugbank_approved_percentile', 'Carcinogens_Lagunin_drugbank_approved_percentile', 'ClinTox_drugbank_approved_percentile', 'DILI_drugbank_approved_percentile', 'HIA_Hou_drugbank_approved_percentile', 'NR-AR-LBD_drugbank_approved_percentile', 'NR-AR_drugbank_approved_percentile', 'NR-AhR_drugbank_approved_percentile', 'NR-Aromatase_drugbank_approved_percentile', 'NR-ER-LBD_drugbank_approved_percentile', 'NR-ER_drugbank_approved_percentile', 'NR-PPAR-gamma_drugbank_approved_percentile', 'PAMPA_NCATS_drugbank_approved_percentile', 'Pgp_Broccatelli_drugbank_approved_percentile', 'SR-ARE_drugbank_approved_percentile', 'SR-ATAD5_drugbank_approved_percentile', 'SR-HSE_drugbank_approved_percentile', 'SR-MMP_drugbank_approved_percentile', 'SR-p53_drugbank_approved_percentile', 'Skin_Reaction_drugbank_approved_percentile', 'hERG_drugbank_approved_percentile', 'Caco2_Wang_drugbank_approved_percentile', 'Clearance_Hepatocyte_AZ_drugbank_approved_percentile', 'Clearance_Microsome_AZ_drugbank_approved_percentile', 'Half_Life_Obach_drugbank_approved_percentile', 'HydrationFreeEnergy_FreeSolv_drugbank_approved_percentile', 'LD50_Zhu_drugbank_approved_percentile', 'Lipophilicity_AstraZeneca_drugbank_approved_percentile', 'PPBR_AZ_drugbank_approved_percentile', 'Solubility_AqSolDB_drugbank_approved_percentile', 'VDss_Lombardo_drugbank_approved_percentile']\n", " for prop in properties:\n", " df_in[prop] = admet_predictions[prop].values\n", " # Apply rounding\n", " df_in = df_in.drop(columns=[\"Mol\"])\n", " cols = df_in.columns\n", " df_in[cols] = df_in[cols].round(3)\n", "\n", " return df_in\n", "\n", "\n", "def process_and_save_molecules(input_file, output_file):\n", " # Load input file\n", " df = (\n", " pd.read_csv(input_file)\n", " if input_file.endswith(\".csv\")\n", " else pd.read_parquet(input_file)\n", " )\n", "\n", " # Add molecular properties\n", " df = add_property_columns(df)\n", "\n", " # Save output file\n", " if output_file.endswith(\".csv\"):\n", " df.to_csv(output_file, index=False)\n", " elif output_file.endswith(\".parquet\"):\n", " df.to_parquet(output_file, index=False)\n", " else:\n", " raise ValueError(\"Output file must be .csv or .parquet\")\n", "\n", " print(f\"Processed DataFrame saved to: {output_file}\")\n", " \n", "# process_and_save_molecules('/lfs/skampere1/0/sttruong/cheapvs_llm/lig_llm.csv', '/lfs/skampere1/0/sttruong/cheapvs_llm/lig_llm_prop.csv')" ] }, { "cell_type": "code", "execution_count": 12, "id": "3c07c1c9", "metadata": {}, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 6/6 [00:00<00:00, 795.48it/s]\n", "Computing physchem properties: 100%|██████████| 6/6 [00:00<00:00, 427.16it/s]\n", "RDKit fingerprints: 100%|██████████| 6/6 [00:00<00:00, 31.16it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.12it/s]\n", "model ensembles: 50%|█████ | 1/2 [00:04<00:04, 4.47s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.10it/s]\n", "model ensembles: 100%|██████████| 2/2 [00:09<00:00, 4.51s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 16/16 [00:00<00:00, 1433.46it/s]\n", "Computing physchem properties: 100%|██████████| 16/16 [00:00<00:00, 658.78it/s]\n", "RDKit fingerprints: 100%|██████████| 16/16 [00:00<00:00, 31.79it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:06<00:00, 1.25s/it]\n", "model ensembles: 50%|█████ | 1/2 [00:06<00:06, 6.24s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:06<00:00, 1.24s/it]\n", "model ensembles: 100%|██████████| 2/2 [00:12<00:00, 6.22s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 15/15 [00:00<00:00, 2246.15it/s]\n", "Computing physchem properties: 100%|██████████| 15/15 [00:00<00:00, 850.10it/s]\n", "RDKit fingerprints: 100%|██████████| 15/15 [00:00<00:00, 35.17it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.07it/s]\n", "model ensembles: 50%|█████ | 1/2 [00:04<00:04, 4.67s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.12it/s]\n", "model ensembles: 100%|██████████| 2/2 [00:09<00:00, 4.56s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 16/16 [00:00<00:00, 5747.59it/s]\n", "Computing physchem properties: 100%|██████████| 16/16 [00:00<00:00, 899.64it/s]\n", "RDKit fingerprints: 100%|██████████| 16/16 [00:00<00:00, 36.56it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:05<00:00, 1.15s/it]\n", "model ensembles: 50%|█████ | 1/2 [00:05<00:05, 5.75s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:06<00:00, 1.26s/it]\n", "model ensembles: 100%|██████████| 2/2 [00:12<00:00, 6.02s/it]\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 27/27 [00:00<00:00, 5983.00it/s]\n", "Computing physchem properties: 100%|██████████| 27/27 [00:00<00:00, 830.19it/s]\n", "RDKit fingerprints: 100%|██████████| 27/27 [00:00<00:00, 36.61it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.09it/s]\n", "model ensembles: 50%|█████ | 1/2 [00:04<00:04, 4.58s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.04it/s]\n", "model ensembles: 100%|██████████| 2/2 [00:09<00:00, 4.69s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 12/12 [00:00<00:00, 2670.26it/s]\n", "Computing physchem properties: 100%|██████████| 12/12 [00:00<00:00, 658.23it/s]\n", "RDKit fingerprints: 100%|██████████| 12/12 [00:00<00:00, 34.39it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.14it/s]\n", "model ensembles: 50%|█████ | 1/2 [00:04<00:04, 4.40s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.19it/s]\n", "model ensembles: 100%|██████████| 2/2 [00:08<00:00, 4.30s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 10/10 [00:00<00:00, 2838.59it/s]\n", "Computing physchem properties: 100%|██████████| 10/10 [00:00<00:00, 1053.13it/s]\n", "RDKit fingerprints: 100%|██████████| 10/10 [00:00<00:00, 38.98it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.14it/s]\n", "model ensembles: 50%|█████ | 1/2 [00:04<00:04, 4.40s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.21it/s]\n", "model ensembles: 100%|██████████| 2/2 [00:08<00:00, 4.27s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 12/12 [00:00<00:00, 6009.03it/s]\n", "Computing physchem properties: 100%|██████████| 12/12 [00:00<00:00, 743.33it/s]\n", "RDKit fingerprints: 100%|██████████| 12/12 [00:00<00:00, 35.72it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.20it/s]\n", "model ensembles: 50%|█████ | 1/2 [00:04<00:04, 4.16s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.07it/s]\n", "model ensembles: 100%|██████████| 2/2 [00:08<00:00, 4.41s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 9/9 [00:00<00:00, 2945.21it/s]\n", "Computing physchem properties: 100%|██████████| 9/9 [00:00<00:00, 583.64it/s]\n", "RDKit fingerprints: 100%|██████████| 9/9 [00:00<00:00, 34.63it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:06<00:00, 1.21s/it]\n", "model ensembles: 50%|█████ | 1/2 [00:06<00:06, 6.03s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:05<00:00, 1.17s/it]\n", "model ensembles: 100%|██████████| 2/2 [00:11<00:00, 5.95s/it]\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 7/7 [00:00<00:00, 2793.28it/s]\n", "Computing physchem properties: 100%|██████████| 7/7 [00:00<00:00, 657.16it/s]\n", "RDKit fingerprints: 100%|██████████| 7/7 [00:00<00:00, 38.13it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.18it/s]\n", "model ensembles: 50%|█████ | 1/2 [00:04<00:04, 4.25s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.20it/s]\n", "model ensembles: 100%|██████████| 2/2 [00:08<00:00, 4.22s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 8/8 [00:00<00:00, 2693.84it/s]\n", "Computing physchem properties: 100%|██████████| 8/8 [00:00<00:00, 702.80it/s]\n", "RDKit fingerprints: 100%|██████████| 8/8 [00:00<00:00, 32.30it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.19it/s]\n", "model ensembles: 50%|█████ | 1/2 [00:04<00:04, 4.22s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.21it/s]\n", "model ensembles: 100%|██████████| 2/2 [00:08<00:00, 4.18s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 14/14 [00:00<00:00, 4036.59it/s]\n", "Computing physchem properties: 100%|██████████| 14/14 [00:00<00:00, 780.49it/s]\n", "RDKit fingerprints: 100%|██████████| 14/14 [00:00<00:00, 34.74it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:05<00:00, 1.15s/it]\n", "model ensembles: 50%|█████ | 1/2 [00:05<00:05, 5.77s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:05<00:00, 1.08s/it]\n", "model ensembles: 100%|██████████| 2/2 [00:11<00:00, 5.59s/it]\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 33/33 [00:00<00:00, 3470.97it/s]\n", "Computing physchem properties: 100%|██████████| 33/33 [00:00<00:00, 664.33it/s]\n", "RDKit fingerprints: 100%|██████████| 33/33 [00:01<00:00, 32.34it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.11it/s]\n", "model ensembles: 50%|█████ | 1/2 [00:04<00:04, 4.51s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.11it/s]\n", "model ensembles: 100%|██████████| 2/2 [00:09<00:00, 4.50s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 8/8 [00:00<00:00, 1897.34it/s]\n", "Computing physchem properties: 100%|██████████| 8/8 [00:00<00:00, 463.09it/s]\n", "RDKit fingerprints: 100%|██████████| 8/8 [00:00<00:00, 23.32it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:05<00:00, 1.11s/it]\n", "model ensembles: 50%|█████ | 1/2 [00:05<00:05, 5.55s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:05<00:00, 1.17s/it]\n", "model ensembles: 100%|██████████| 2/2 [00:11<00:00, 5.70s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 58/58 [00:00<00:00, 5965.12it/s]\n", "Computing physchem properties: 100%|██████████| 58/58 [00:00<00:00, 825.65it/s]\n", "RDKit fingerprints: 100%|██████████| 58/58 [00:01<00:00, 34.79it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.04it/s]\n", "model ensembles: 50%|█████ | 1/2 [00:04<00:04, 4.82s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.08it/s]\n", "model ensembles: 100%|██████████| 2/2 [00:09<00:00, 4.72s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 13/13 [00:00<00:00, 3550.56it/s]\n", "Computing physchem properties: 100%|██████████| 13/13 [00:00<00:00, 892.67it/s]\n", "RDKit fingerprints: 100%|██████████| 13/13 [00:00<00:00, 38.87it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:05<00:00, 1.10s/it]\n", "model ensembles: 50%|█████ | 1/2 [00:05<00:05, 5.51s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:05<00:00, 1.12s/it]\n", "model ensembles: 100%|██████████| 2/2 [00:11<00:00, 5.57s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 6/6 [00:00<00:00, 1764.41it/s]\n", "Computing physchem properties: 100%|██████████| 6/6 [00:00<00:00, 564.02it/s]\n", "RDKit fingerprints: 100%|██████████| 6/6 [00:00<00:00, 30.95it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:03<00:00, 1.26it/s]\n", "model ensembles: 50%|█████ | 1/2 [00:03<00:03, 3.96s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.20it/s]\n", "model ensembles: 100%|██████████| 2/2 [00:08<00:00, 4.06s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 5/5 [00:00<00:00, 1885.76it/s]\n", "Computing physchem properties: 100%|██████████| 5/5 [00:00<00:00, 474.87it/s]\n", "RDKit fingerprints: 100%|██████████| 5/5 [00:00<00:00, 31.64it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:03<00:00, 1.26it/s]\n", "model ensembles: 50%|█████ | 1/2 [00:03<00:03, 3.96s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:03<00:00, 1.26it/s]\n", "model ensembles: 100%|██████████| 2/2 [00:07<00:00, 3.97s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 20/20 [00:00<00:00, 4126.83it/s]\n", "Computing physchem properties: 100%|██████████| 20/20 [00:00<00:00, 776.83it/s]\n", "RDKit fingerprints: 100%|██████████| 20/20 [00:00<00:00, 33.30it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:05<00:00, 1.09s/it]\n", "model ensembles: 50%|█████ | 1/2 [00:05<00:05, 5.48s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.08it/s]\n", "model ensembles: 100%|██████████| 2/2 [00:10<00:00, 5.06s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 26/26 [00:00<00:00, 4934.48it/s]\n", "Computing physchem properties: 100%|██████████| 26/26 [00:00<00:00, 954.84it/s]\n", "RDKit fingerprints: 100%|██████████| 26/26 [00:00<00:00, 37.89it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.14it/s]\n", "model ensembles: 50%|█████ | 1/2 [00:04<00:04, 4.41s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.23it/s]\n", "model ensembles: 100%|██████████| 2/2 [00:08<00:00, 4.24s/it]\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:473: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " vars(torch.load(path, map_location=lambda storage, loc: storage)[\"args\"]),\n", "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:112: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "INFO: Pandarallel will run on 12 workers.\n", "INFO: Pandarallel will use Memory file system to transfer data between the main process and workers.\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "/lfs/local/0/sttruong/miniconda3/envs/cheapvs_llm/lib/python3.10/site-packages/chemprop/utils.py:418: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n", " state = torch.load(path, map_location=lambda storage, loc: storage)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n", "Loading pretrained parameter \"encoder.encoder.0.cached_zero_vector\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_i.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_h.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.weight\".\n", "Loading pretrained parameter \"encoder.encoder.0.W_o.bias\".\n", "Loading pretrained parameter \"readout.1.weight\".\n", "Loading pretrained parameter \"readout.1.bias\".\n", "Loading pretrained parameter \"readout.4.weight\".\n", "Loading pretrained parameter \"readout.4.bias\".\n", "Moving model to cuda\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "SMILES to Mol: 100%|██████████| 6/6 [00:00<00:00, 3400.79it/s]\n", "Computing physchem properties: 100%|██████████| 6/6 [00:00<00:00, 569.30it/s]\n", "RDKit fingerprints: 100%|██████████| 6/6 [00:00<00:00, 32.30it/s]\n", "model ensembles: 0%| | 0/2 [00:00, ?it/s]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:04<00:00, 1.02it/s]\n", "model ensembles: 50%|█████ | 1/2 [00:04<00:04, 4.91s/it]\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "\u001b[A\n", "individual models: 100%|██████████| 5/5 [00:05<00:00, 1.05s/it]\n", "model ensembles: 100%|██████████| 2/2 [00:10<00:00, 5.09s/it]\n" ] } ], "source": [ "import pandas as pd\n", "\n", "files = os.listdir('/lfs/skampere1/0/sttruong/cheapvs_llm/20_targets/drugs_filtered/')\n", "for file in files:\n", " if file.endswith('.csv'):\n", " df = pd.read_csv(f'/lfs/skampere1/0/sttruong/cheapvs_llm/20_targets/drugs_filtered/{file}')\n", " df = add_property_columns(df)\n", " name = file.split('.')[0] + '_prop.csv'\n", " df.to_csv(f'/lfs/skampere1/0/sttruong/cheapvs_llm/20_targets/drugs_filtered/{file}', index=False)" ] }, { "cell_type": "code", "execution_count": null, "id": "af944524", "metadata": {}, "outputs": [], "source": [ "import pandas as pd\n", "\n", "df_20k = pd.read_csv('lig_llm_prop.csv')\n", "df_drug = pd.read_csv('20_targets/drugs_filtered/PDE4A_HUMAN.csv')\n", "combined_df = pd.concat([df_20k, df_drug], ignore_index=True)\n" ] }, { "cell_type": "code", "execution_count": 23, "id": "b7729b6e", "metadata": {}, "outputs": [ { "data": { "text/html": [ "
| \n", " | SMILES | \n", "MW | \n", "LogP | \n", "numHDonors | \n", "numHAcceptors | \n", "TPSA | \n", "rotBonds | \n", "Arom | \n", "QED | \n", "SA | \n", "... | \n", "Half_Life_Obach_drugbank_approved_percentile | \n", "HydrationFreeEnergy_FreeSolv_drugbank_approved_percentile | \n", "LD50_Zhu_drugbank_approved_percentile | \n", "Lipophilicity_AstraZeneca_drugbank_approved_percentile | \n", "PPBR_AZ_drugbank_approved_percentile | \n", "Solubility_AqSolDB_drugbank_approved_percentile | \n", "VDss_Lombardo_drugbank_approved_percentile | \n", "DrugBank ID | \n", "Name | \n", "Drug Groups | \n", "
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| 0 | \n", "COc1cc(C=Cc2nc(-c3ccccc3)cs2)ccc1OCC#N | \n", "348.427 | \n", "4.891 | \n", "0 | \n", "5 | \n", "55.14 | \n", "6 | \n", "3 | \n", "0.636 | \n", "2.317 | \n", "... | \n", "15.432 | \n", "61.303 | \n", "38.038 | \n", "95.812 | \n", "92.555 | \n", "2.792 | \n", "4.769 | \n", "NaN | \n", "NaN | \n", "NaN | \n", "
| 1 | \n", "COc1ccn2c(C=C(C#N)C(=O)c3ccc(Cl)cc3F)cnc2c1 | \n", "355.756 | \n", "3.925 | \n", "0 | \n", "5 | \n", "67.39 | \n", "4 | \n", "3 | \n", "0.404 | \n", "2.755 | \n", "... | \n", "62.389 | \n", "38.813 | \n", "84.878 | \n", "85.770 | \n", "87.864 | \n", "4.886 | \n", "84.839 | \n", "NaN | \n", "NaN | \n", "NaN | \n", "
| 2 | \n", "C#CCOc1ccc(C=C(Cl)c2nc3ccc(C)cc3o2)cc1OC | \n", "353.805 | \n", "4.894 | \n", "0 | \n", "4 | \n", "44.49 | \n", "5 | \n", "3 | \n", "0.612 | \n", "2.612 | \n", "... | \n", "79.643 | \n", "81.078 | \n", "32.765 | \n", "98.914 | \n", "92.206 | \n", "3.877 | \n", "96.472 | \n", "NaN | \n", "NaN | \n", "NaN | \n", "
| 3 | \n", "CC(Sc1nnc(-c2ccccc2Cl)n1N)c1ncc(C(C)(C)C)o1 | \n", "377.901 | \n", "4.451 | \n", "1 | \n", "7 | \n", "82.76 | \n", "4 | \n", "3 | \n", "0.534 | \n", "3.167 | \n", "... | \n", "15.859 | \n", "65.762 | \n", "26.289 | \n", "93.292 | \n", "80.264 | \n", "17.914 | \n", "90.810 | \n", "NaN | \n", "NaN | \n", "NaN | \n", "
| 4 | \n", "NC(=O)c1ccccc1OCC(=O)Nc1ccc(OC(F)F)c(Cl)c1 | \n", "370.739 | \n", "3.058 | \n", "2 | \n", "4 | \n", "90.65 | \n", "7 | \n", "2 | \n", "0.783 | \n", "1.962 | \n", "... | \n", "63.358 | \n", "35.169 | \n", "25.514 | \n", "78.635 | \n", "64.754 | \n", "10.275 | \n", "5.506 | \n", "NaN | \n", "NaN | \n", "NaN | \n", "
| ... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "... | \n", "
| 20364 | \n", "COC1=CC=CC(NC2=C(C=NC3=C(C)C=C(C=C23)S(=O)(=O)... | \n", "518.595 | \n", "3.929 | \n", "2 | \n", "7 | \n", "131.69 | \n", "7 | \n", "4 | \n", "0.379 | \n", "2.571 | \n", "... | \n", "96.859 | \n", "18.457 | \n", "60.605 | \n", "72.664 | \n", "65.568 | \n", "15.006 | \n", "12.214 | \n", "DB12137 | \n", "GSK-256066 | \n", "investigational | \n", "
| 20365 | \n", "CS(=O)(=O)NC1=CC2=C(OC3=C2C(=CC=C3OC(F)F)C(=O)... | \n", "516.309 | \n", "5.513 | \n", "2 | \n", "6 | \n", "110.53 | \n", "6 | \n", "4 | \n", "0.350 | \n", "2.774 | \n", "... | \n", "99.612 | \n", "26.871 | \n", "80.109 | \n", "69.290 | \n", "86.933 | \n", "6.088 | \n", "91.663 | \n", "DB12375 | \n", "Oglemilast | \n", "investigational | \n", "
| 20366 | \n", "[O-][N+]1=CC(=CC=C1)C#CC1=CC(=CC=C1)N1C=C(C(=O... | \n", "422.444 | \n", "2.311 | \n", "1 | \n", "5 | \n", "90.93 | \n", "3 | \n", "4 | \n", "0.312 | \n", "2.887 | \n", "... | \n", "50.834 | \n", "34.122 | \n", "79.953 | \n", "63.978 | \n", "70.492 | \n", "31.214 | \n", "29.508 | \n", "DB13029 | \n", "MK-0873 | \n", "investigational | \n", "
| 20367 | \n", "COC1=CC=C(C=C1OC1CCCC1)C(CC(N)=O)N1C(=O)C2=C(C... | \n", "408.454 | \n", "3.229 | \n", "1 | \n", "5 | \n", "98.93 | \n", "7 | \n", "2 | \n", "0.709 | \n", "2.776 | \n", "... | \n", "37.379 | \n", "60.256 | \n", "51.105 | \n", "68.205 | \n", "74.564 | \n", "7.445 | \n", "88.212 | \n", "DB15640 | \n", "CDC-801 | \n", "investigational | \n", "
| 20368 | \n", "OB1OCC2=C1C=CC(OC1=CC(C#N)=C(C=C1)C#N)=C2 | \n", "276.060 | \n", "1.440 | \n", "1 | \n", "5 | \n", "86.27 | \n", "2 | \n", "2 | \n", "0.838 | \n", "2.894 | \n", "... | \n", "10.237 | \n", "54.323 | \n", "91.121 | \n", "52.734 | \n", "73.052 | \n", "22.955 | \n", "24.157 | \n", "DB16039 | \n", "AN2898 | \n", "investigational | \n", "
20369 rows × 101 columns
\n", "