You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/example_pipeline.rst
+58-2Lines changed: 58 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -2,6 +2,8 @@
2
2
Example pipeline setup and execution
3
3
=====================================
4
4
5
+
.. _common-setup:
6
+
5
7
Preliminary setup and common instructions
6
8
-----------------------------------------
7
9
@@ -14,7 +16,7 @@ Before execution of the ``MiCoNE`` pipeline we need to install the environments:
14
16
15
17
micone install
16
18
17
-
.. note:: This command will take a considerable amount of time (several hours) as MiCoNE will install all the ``conda`` environments
19
+
.. warning:: This command will take a considerable amount of time (several hours) as MiCoNE will install all the ``conda`` environments
18
20
19
21
If you wish to install only a subset of the environments, you can specify the environments to install using the ``-e`` option:
20
22
@@ -67,8 +69,62 @@ To run the pipeline, you can use the ``run.sh`` script in the ``<pipeline_dir>``
67
69
Full pipeline workflow
68
70
----------------------
69
71
72
+
First follow the instructions in steps 1-3 in the :ref:`common-setup` section.
73
+
74
+
Let us assume that you have multiplexed (``run1``, ``run2``, and ``run3``) paired end 16S sequence data stored in the ``<pipeline_dir>/seqs`` folder. To run the pipeline you will need the following:
75
+
76
+
1. ``forward.fastq.gz``: Forward reads
77
+
2. ``reverse.fastq.gz``: Reverse reads
78
+
3. ``barcodes.fastq.gz``: Barcodes
79
+
4. ``mapping.tsv``: Mapping file
80
+
5. ``sample_metadata.tsv``: Sample metadata file
81
+
82
+
.. warning:: Keep the file names as they are. The pipeline might have issues if the file names are changed.
83
+
84
+
An example ``sample_sheet.csv`` file will look like this:
.. note:: These files must follow the ``qiime2`` supported formats. For more information about the supported formats see the `qiime2 documentation <https://docs.qiime2.org/2023.7/tutorials/importing/>`__.
97
+
98
+
70
99
Network inference workflow
71
100
---------------------------
72
101
73
-
Note that this workflow is only valid if your biom files already have taxonomy labels assigned. You must run the workflow from the TA step if they do not.
102
+
Before running this workflow make sure that your OTU tables have taxonomy metadata and sample metadata information.
103
+
You must run the workflow from the TA step if they do not.
104
+
105
+
First follow the instructions in steps 1-3 in the :ref:`common-setup` section.
106
+
107
+
Let us assume that you have 3 sets of OTU tables (``id1``, ``id2``, and ``id3``) you wish to analyze. To run the pipeline you will need the following:
108
+
109
+
1. ``otu_table.tsv``: OTU table in ``.tsv`` format
110
+
2. ``obs_metadata.tsv``: Taxonomy assignments in ``.tsv`` format. It must contain the following columns: "Kingdom", "Phylum", "Class", "Order", "Family", "Genus", "Species". The latter columns can be dropped if you have grouped your taxonomy at a higher level.
111
+
3. ``sample_metadata.tsv``: Sample metadata file
112
+
4. ``children_map.json``: A file that maps the current taxonomy ids to lower taxonomic level. Can be an empty JSON if you wish to ignore this field.
113
+
114
+
.. warning:: Keep the file names as they are. The pipeline might have issues if the file names are changed.
115
+
116
+
An example ``sample_sheet.csv`` file will look like this:
0 commit comments