diff --git a/sdk/src/account/mod.rs b/sdk/src/account/mod.rs index 5a21636..63d2718 100644 --- a/sdk/src/account/mod.rs +++ b/sdk/src/account/mod.rs @@ -117,7 +117,7 @@ impl Account { fn __hash__(&self) -> u64 { let mut hasher = DefaultHasher::new(); - // because it's enouth to hash the private key only we add a dummy string so that: + // because it's enough to hash the private key only we add a dummy string so that: // hash(Account) != hash(PrivateKey) "account".hash(&mut hasher); self.private_key.hash(&mut hasher); diff --git a/zkml-research/MNIST_preprocessing/6_presentation.ipynb b/zkml-research/MNIST_preprocessing/6_presentation.ipynb index 2e3958d..4776fa9 100644 --- a/zkml-research/MNIST_preprocessing/6_presentation.ipynb +++ b/zkml-research/MNIST_preprocessing/6_presentation.ipynb @@ -167,7 +167,7 @@ "metadata": {}, "outputs": [], "source": [ - "# conver to pytorch tensors\n", + "# convertto pytorch tensors\n", "import torch\n", "\n", "train_images_tensor_initial = torch.from_numpy(train_images_2d).float()\n", diff --git a/zkml-research/MNIST_preprocessing/7_trees.ipynb b/zkml-research/MNIST_preprocessing/7_trees.ipynb index 8b9978a..fbfcb32 100644 --- a/zkml-research/MNIST_preprocessing/7_trees.ipynb +++ b/zkml-research/MNIST_preprocessing/7_trees.ipynb @@ -167,7 +167,7 @@ "metadata": {}, "outputs": [], "source": [ - "# conver to pytorch tensors\n", + "# convertto pytorch tensors\n", "import torch\n", "\n", "train_images_tensor_initial = torch.from_numpy(train_images_2d).float()\n", diff --git a/zkml-research/MNIST_preprocessing/8_further_preprocessing.ipynb b/zkml-research/MNIST_preprocessing/8_further_preprocessing.ipynb index 4a49596..239f96f 100644 --- a/zkml-research/MNIST_preprocessing/8_further_preprocessing.ipynb +++ b/zkml-research/MNIST_preprocessing/8_further_preprocessing.ipynb @@ -167,7 +167,7 @@ "metadata": {}, "outputs": [], "source": [ - "# conver to pytorch tensors\n", + "# convertto pytorch tensors\n", "import torch\n", "\n", "train_images_tensor_initial = torch.from_numpy(train_images_2d).float()\n", diff --git a/zkml-research/zkml-web-app/mnist_train.ipynb b/zkml-research/zkml-web-app/mnist_train.ipynb index c735e4a..74fe8a9 100644 --- a/zkml-research/zkml-web-app/mnist_train.ipynb +++ b/zkml-research/zkml-web-app/mnist_train.ipynb @@ -13,7 +13,7 @@ "source": [ "MLP neural networks are expressive ML models. Here, we use them for the MNIST classification task, which contains images of handwritten digits. We show the inference of MLP neural networks for the MNIST dataset is possible in a zero knowledge environment.\n", "\n", - "For this, we first download the dataset, and then compute feature representations of the dataset. We then train and test an MLP neural network on the feature dataset using PyTorch (beacuse we can train with L1 regularization which is helpful for pruning later on). Afterward, we iteratively prune the network (meaning we set weights and biases close to 0 to actually 0, this will save cicruit constraints later on) and fine-tune it. Then, we convert the final PyTorch MLP model to a scikit-learn MLP model (since the transpiler supports scikit-learn models) and transpile the final MLP neural network to Leo, evaluate the Leo network and create a zero knowledge proof." + "For this, we first download the dataset, and then compute feature representations of the dataset. We then train and test an MLP neural network on the feature dataset using PyTorch (because we can train with L1 regularization which is helpful for pruning later on). Afterward, we iteratively prune the network (meaning we set weights and biases close to 0 to actually 0, this will save circuitconstraints later on) and fine-tune it. Then, we convert the final PyTorch MLP model to a scikit-learn MLP model (since the transpiler supports scikit-learn models) and transpile the final MLP neural network to Leo, evaluate the Leo network and create a zero knowledge proof." ] }, { @@ -235,7 +235,7 @@ "metadata": {}, "outputs": [], "source": [ - "# conver to pytorch tensors\n", + "# convertto pytorch tensors\n", "import torch\n", "\n", "train_images_tensor_initial = torch.from_numpy(train_images).float()\n", diff --git a/zkml/examples/MLP_neural_networks/mnist_dataset.ipynb b/zkml/examples/MLP_neural_networks/mnist_dataset.ipynb index fc5e3c2..30e7731 100644 --- a/zkml/examples/MLP_neural_networks/mnist_dataset.ipynb +++ b/zkml/examples/MLP_neural_networks/mnist_dataset.ipynb @@ -13,7 +13,7 @@ "source": [ "MLP neural networks are expressive ML models. Here, we use them for the MNIST classification task, which contains images of handwritten digits. We show the inference of MLP neural networks for the MNIST dataset is possible in a zero knowledge environment.\n", "\n", - "For this, we first download the dataset, and then compute feature representations of the dataset. We then train and test an MLP neural network on the feature dataset using PyTorch (beacuse we can train with L1 regularization which is helpful for pruning later on). Afterward, we iteratively prune the network (meaning we set weights and biases close to 0 to actually 0, this will save cicruit constraints later on) and fine-tune it. Then, we convert the final PyTorch MLP model to a scikit-learn MLP model (since the transpiler supports scikit-learn models) and transpile the final MLP neural network to Leo, evaluate the Leo network and create a zero knowledge proof." + "For this, we first download the dataset, and then compute feature representations of the dataset. We then train and test an MLP neural network on the feature dataset using PyTorch (because we can train with L1 regularization which is helpful for pruning later on). Afterward, we iteratively prune the network (meaning we set weights and biases close to 0 to actually 0, this will save circuitconstraints later on) and fine-tune it. Then, we convert the final PyTorch MLP model to a scikit-learn MLP model (since the transpiler supports scikit-learn models) and transpile the final MLP neural network to Leo, evaluate the Leo network and create a zero knowledge proof." ] }, { @@ -235,7 +235,7 @@ "metadata": {}, "outputs": [], "source": [ - "# conver to pytorch tensors\n", + "# convertto pytorch tensors\n", "import torch\n", "\n", "train_images_tensor_initial = torch.from_numpy(train_images).float()\n", diff --git a/zkml/tests/helper.py b/zkml/tests/helper.py index 4c1cab8..6c8aa0e 100644 --- a/zkml/tests/helper.py +++ b/zkml/tests/helper.py @@ -271,7 +271,7 @@ def read_idx1_ubyte_label_file(filename): - # conver to pytorch tensors + # convertto pytorch tensors import torch train_images_tensor_initial = torch.from_numpy(train_images_2d).float()