|
33 | 33 | " * Mambaforge (a package manager for bioinformatics tools).\n", |
34 | 34 | " * `sra-tools`, `perl-dbd-sqlite`, and `perl-dbi` (specific bioinformatics packages).\n", |
35 | 35 | " * Nextflow (a workflow management system).\n", |
36 | | - " * `gsutil` (for interacting with Google Cloud Storage).\n", |
| 36 | + " * `aws s3` (for interacting with AWS S3 Storage).\n", |
37 | 37 | "\n", |
38 | | - "3. **Download and organize necessary data:** Students will download the TransPi transcriptome assembly software and its associated resources (databases, scripts, configuration files) from a Google Cloud Storage bucket. This includes understanding the directory structure and file organization.\n", |
| 38 | + "3. **Download and organize necessary data:** Students will download the TransPi transcriptome assembly software and its associated resources (databases, scripts, configuration files) from an S3 bucket. This includes understanding the directory structure and file organization.\n", |
39 | 39 | "\n", |
40 | 40 | "4. **Manage file permissions:** Students will use the `chmod` command to set executable permissions for the necessary files and directories within the TransPi software.\n", |
41 | 41 | "\n", |
|
53 | 53 | "* **Shell Access:** The ability to execute shell commands from within the Jupyter Notebook environment (using `!` and `%`).\n", |
54 | 54 | "* **Java Development Kit (JDK):** Required for Nextflow.\n", |
55 | 55 | "* **Miniforge** A package manager for installing bioinformatics tools.\n", |
56 | | - "* **`gsutil`:** The Google Cloud Storage command-line tool. This is crucial for downloading data from Google Cloud Storage." |
| 56 | + "* **`aws s3`:** The AWS command-line tool. This is crucial for downloading data from an S3 storage bucket." |
57 | 57 | ] |
58 | 58 | }, |
59 | 59 | { |
|
104 | 104 | "## Time to begin!\n", |
105 | 105 | "\n", |
106 | 106 | "**Step 1:** To start, make sure that you are in the right starting place with a `cd`.\n", |
107 | | - "> `pwd` prints our current local working directory. Make sure the output from the command is: `/home/jupyter`" |
| 107 | + "> `pwd` prints our current local working directory. Make sure the output from the command is: `/home/ec2-user/SageMaker`" |
108 | 108 | ] |
109 | 109 | }, |
110 | 110 | { |
|
114 | 114 | "metadata": {}, |
115 | 115 | "outputs": [], |
116 | 116 | "source": [ |
117 | | - "%cd /home/jupyter" |
| 117 | + "%cd /home/ec2-user/SageMaker" |
118 | 118 | ] |
119 | 119 | }, |
120 | 120 | { |
|
147 | 147 | "! java -version" |
148 | 148 | ] |
149 | 149 | }, |
150 | | - { |
151 | | - "cell_type": "markdown", |
152 | | - "id": "7b3ffb16-3395-4c01-9774-ee568e815490", |
153 | | - "metadata": {}, |
154 | | - "source": [ |
155 | | - "**Step 3:** Install Miniforge (a package manager), which is needed to support the information held within the TransPi databases." |
156 | | - ] |
157 | | - }, |
158 | | - { |
159 | | - "cell_type": "code", |
160 | | - "execution_count": null, |
161 | | - "id": "ac5b204a-f0db-4ceb-bf37-57eca6d77974", |
162 | | - "metadata": {}, |
163 | | - "outputs": [], |
164 | | - "source": [ |
165 | | - "! curl -L -O https://github.com/conda-forge/miniforge/releases/latest/download/Miniforge3-$(uname)-$(uname -m).sh\n", |
166 | | - "! bash Miniforge3-$(uname)-$(uname -m).sh -b -p $HOME/miniforge" |
167 | | - ] |
168 | | - }, |
169 | | - { |
170 | | - "cell_type": "markdown", |
171 | | - "id": "c5584e2e", |
172 | | - "metadata": {}, |
173 | | - "source": [ |
174 | | - "Next, add it to the path." |
175 | | - ] |
176 | | - }, |
177 | | - { |
178 | | - "cell_type": "code", |
179 | | - "execution_count": null, |
180 | | - "id": "ad030cd1", |
181 | | - "metadata": {}, |
182 | | - "outputs": [], |
183 | | - "source": [ |
184 | | - "import os\n", |
185 | | - "os.environ[\"PATH\"] += os.pathsep + os.environ[\"HOME\"]+\"/miniforge/bin\"" |
186 | | - ] |
187 | | - }, |
188 | 150 | { |
189 | 151 | "cell_type": "markdown", |
190 | 152 | "id": "7b930ad7", |
191 | 153 | "metadata": {}, |
192 | 154 | "source": [ |
193 | | - "Next, using Miniforge and bioconda, install the tools that will be used in this tutorial." |
| 155 | + "**Step 3:** Using Mamba and bioconda, install the tools that will be used in this tutorial." |
194 | 156 | ] |
195 | 157 | }, |
196 | 158 | { |
|
239 | 201 | "metadata": {}, |
240 | 202 | "outputs": [], |
241 | 203 | "source": [ |
242 | | - "! gsutil -m cp -r gs://nigms-sandbox/nosi-inbremaine-storage/TransPi ./" |
| 204 | + "! aws s3 cp --recursive s3://nigms-sandbox/nosi-inbremaine-storage/TransPi ./TransPi" |
243 | 205 | ] |
244 | 206 | }, |
245 | 207 | { |
|
249 | 211 | "source": [ |
250 | 212 | "<div class=\"alert alert-block alert-success\">\n", |
251 | 213 | " <i class=\"fa fa-hand-paper-o\" aria-hidden=\"true\"></i>\n", |
252 | | - " <b>Note: </b> gsutil\n", |
| 214 | + " <b>Note: </b> aws\n", |
253 | 215 | "</div>\n", |
254 | 216 | "\n", |
255 | | - ">`gsutil` is a tool allows you to interact with Google Cloud Storage through the command line." |
| 217 | + ">`aws s3` is a tool allows you to interact with S3 Storage through the command line." |
256 | 218 | ] |
257 | 219 | }, |
258 | 220 | { |
|
277 | 239 | "metadata": {}, |
278 | 240 | "outputs": [], |
279 | 241 | "source": [ |
280 | | - "! gsutil -m cp -r gs://nigms-sandbox/nosi-inbremaine-storage/resources ./" |
| 242 | + "! aws s3 cp --recursive s3://nigms-sandbox/nosi-inbremaine-storage/resources ./resources" |
281 | 243 | ] |
282 | 244 | }, |
283 | 245 | { |
|
302 | 264 | "> - They can also be stacked so `../../` will take you two layers up.\n", |
303 | 265 | ">\n", |
304 | 266 | ">- If you were to type `!ls ./nextWeek/` it would return the contents of the `nextWeek` directory which is one layer down from the current directory, so it would return `manyThings.txt`.\n", |
305 | | - ">\n", |
306 | | - ">**This means that in the second line of the code cell above, the file `TransPi.nf` will be copied from the Google Cloud Storage bucket to the current directory.**" |
| 267 | + ">" |
307 | 268 | ] |
308 | 269 | }, |
309 | 270 | { |
|
377 | 338 | "outputs": [], |
378 | 339 | "source": [ |
379 | 340 | "from jupyterquiz import display_quiz\n", |
380 | | - "display_quiz(\"../quiz-material/01-cp1.json\", shuffle_questions = True)" |
| 341 | + "display_quiz(\"Transcriptome-Assembly-Refinement-and-Applications/quiz-material/01-cp1.json\", shuffle_questions = True)" |
381 | 342 | ] |
382 | 343 | }, |
383 | 344 | { |
|
401 | 362 | ] |
402 | 363 | } |
403 | 364 | ], |
404 | | - "metadata": {}, |
| 365 | + "metadata": { |
| 366 | + "kernelspec": { |
| 367 | + "display_name": "conda_python3", |
| 368 | + "language": "python", |
| 369 | + "name": "conda_python3" |
| 370 | + }, |
| 371 | + "language_info": { |
| 372 | + "codemirror_mode": { |
| 373 | + "name": "ipython", |
| 374 | + "version": 3 |
| 375 | + }, |
| 376 | + "file_extension": ".py", |
| 377 | + "mimetype": "text/x-python", |
| 378 | + "name": "python", |
| 379 | + "nbconvert_exporter": "python", |
| 380 | + "pygments_lexer": "ipython3", |
| 381 | + "version": "3.10.16" |
| 382 | + } |
| 383 | + }, |
405 | 384 | "nbformat": 4, |
406 | 385 | "nbformat_minor": 5 |
407 | 386 | } |
0 commit comments