Skip to content

Commit f1413ee

Browse files
committed
Update pip setup and documentation.
1 parent b478af5 commit f1413ee

4 files changed

Lines changed: 39 additions & 16 deletions

File tree

README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -64,9 +64,9 @@ See the [quick start](https://github.com/google/deepconsensus/blob/main/docs/qui
6464

6565
## Where does DeepConsensus fit into my pipeline?
6666

67-
After a PacBio sequencing run, DeepConsensus is meant to be run on the CCS reads
68-
and subreads to create new corrected reads in FASTQ format that can take the
69-
place of the CCS reads for downstream analyses.
67+
After a PacBio sequencing run, DeepConsensus is meant to be run on the subreads
68+
to create new corrected reads in FASTQ format that can take the place of the CCS
69+
reads for downstream analyses.
7070

7171
See the [quick start](https://github.com/google/deepconsensus/blob/main/docs/quick_start.md)
7272
for an example of inputs and outputs.

README_pip.md

Lines changed: 21 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,21 @@
1+
# Important: Pip install is different for CPU versus GPU
2+
3+
If you're on a GPU machine:
4+
5+
```bash
6+
pip install deepconsensus[gpu]==0.2.0
7+
# To make sure the `deepconsensus` CLI works, set the PATH:
8+
export PATH="/home/${USER}/.local/bin:${PATH}"
9+
```
10+
11+
If you're on a CPU machine:
12+
13+
```bash
14+
pip install deepconsensus[cpu]==0.2.0
15+
# To make sure the `deepconsensus` CLI works, set the PATH:
16+
export PATH="/home/${USER}/.local/bin:${PATH}"
17+
```
18+
19+
## Documentation, quick start, citation
20+
21+
All other documentation is on GitHub: [https://github.com/google/deepconsensus](https://github.com/google/deepconsensus).

docs/quick_start.md

Lines changed: 14 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ This covers the following stages:
99
to use DeepConsensus from existing *ccs* reads, but yield will be higher when
1010
including all reads)
1111
2. Aligning subreads to the *ccs* consensus with *[actc]*
12-
3. Running DeepConsensus using one of two options (with pip or using Docker)
12+
3. Running DeepConsensus using either pip or Docker
1313

1414
## System configuration
1515

@@ -24,9 +24,9 @@ GPU: 1 nvidia-tesla-p100
2424
```
2525

2626
DeepConsensus can be run on any compatible Unix systems. In this case, we used a
27-
[n1-standard-16 machine on GCP](https://cloud.google.com/compute/docs/general-purpose-machines#n1_machines), with a NVIDIA P100 GPU.
27+
[n1-standard-16 machine on GCP](https://cloud.google.com/compute/docs/general-purpose-machines#n1_machines), with an NVIDIA P100 GPU.
2828

29-
## Download data for testing
29+
## Download example data
3030

3131
This will download about 142 MB of data and the model is another 245 MB.
3232

@@ -40,16 +40,17 @@ MODEL_DIR="${QUICKSTART_DIRECTORY}/model"
4040
mkdir -p "${DATA}"
4141
mkdir -p "${MODEL_DIR}"
4242

43-
# Download the input data which is PacBio subreads.
43+
# Download the input data, which is PacBio subreads.
4444
gsutil cp gs://brain-genomics-public/research/deepconsensus/quickstart/v0.2/subreads.bam* "${DATA}"/
4545

46-
# Download DeepConsensus model.
46+
# Download the DeepConsensus model.
4747
gsutil cp gs://brain-genomics-public/research/deepconsensus/models/v0.2/* "${MODEL_DIR}"/
4848
```
4949

5050
## If running with GPU, set up your GPU machine correctly.
5151

5252
In our example run, because we're using GPU, we used:
53+
5354
```bash
5455
curl https://raw.githubusercontent.com/google/deepvariant/r1.3/scripts/install_nvidia_docker.sh -o install_nvidia_docker.sh
5556
bash install_nvidia_docker.sh
@@ -62,8 +63,8 @@ to make sure our GPU is set up correctly.
6263
You can install *[ccs]* and *[actc]* on your own. For convenience, we put them in
6364
a Docker image:
6465

65-
```
66-
DOCKER_IMAGE=google/deepconsensus:0.2.0rc1-gpu
66+
```bash
67+
DOCKER_IMAGE=google/deepconsensus:0.2.0-gpu
6768
sudo docker pull ${DOCKER_IMAGE}
6869
```
6970

@@ -84,7 +85,7 @@ quality threshold.
8485
If you want to split up the task for parallelization, we recommend using the
8586
`--chunk` option in *ccs*.
8687

87-
Then, we create `subreads_to_ccs.bam` was created by running *actc*:
88+
Then, we create `subreads_to_ccs.bam` by running *actc*:
8889

8990
```bash
9091
sudo docker run -v "${DATA}":"/data" ${DOCKER_IMAGE} \
@@ -94,7 +95,7 @@ sudo docker run -v "${DATA}":"/data" ${DOCKER_IMAGE} \
9495
/data/subreads_to_ccs.bam
9596
```
9697

97-
DeepConsensus will take FASTA format of *ccs*.
98+
DeepConsensus will take the consensus sequences output by *ccs* in FASTA format.
9899

99100
*actc* already converted the BAM into FASTA. Rename and index it.
100101

@@ -113,7 +114,7 @@ sudo docker run -v "${DATA}":"/data" ${DOCKER_IMAGE} \
113114
You can install DeepConsensus using `pip`:
114115

115116
```bash
116-
pip install deepconsensus[gpu]==0.2.0rc1
117+
pip install deepconsensus[gpu]==0.2.0
117118
```
118119

119120
NOTE: If you're using a CPU machine, install with `deepconsensus[cpu]` instead.
@@ -139,14 +140,15 @@ time deepconsensus run \
139140
```
140141

141142
At the end of your run, you should see:
143+
142144
```
143145
Processed 1000 ZMWs in 341.3297851085663 seconds
144146
Outcome counts: OutcomeCounter(empty_sequence=0, only_gaps_and_padding=50, failed_quality_filter=424, failed_length_filter=0, success=526)
145147
```
146-
the outputs can be found at the following paths:
148+
149+
The final output FASTQ can be found at the following path:
147150

148151
```bash
149-
# Final output fastq file which has DeepConsensus reads.
150152
ls "${DATA}"/output.fastq
151153
```
152154

setup.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@
3838
here = pathlib.Path(__file__).parent.resolve()
3939

4040
# Get the long description from the README file
41-
long_description = (here / 'README.md').read_text(encoding='utf-8')
41+
long_description = (here / 'README_pip.md').read_text(encoding='utf-8')
4242

4343
REQUIREMENTS = (here / 'requirements.txt').read_text().splitlines()
4444
EXTRA_REQUIREMENTS = {

0 commit comments

Comments
 (0)