Cannot download Chocophlan database

Hi,

I have been having an immense amount of trouble being able to download the full_chocophlan.v201901_v31.tar.gz database for humann3. Everything else has gone well, but every time I use the command;
humann_databases --download chocophlan full path/to/directory

I eventually get an error. The thing is that the download actually does progress, but eventually it seems like the system gives up and i get the following error;

CRITICAL ERROR: Unable to download and extract from URL: http://huttenhower.sph.harvard.edu/humann_data/chocophlan/full_chocophlan.v201901_v31.tar.gz

I have tried downloading it directly from the http://huttenhower.sph.harvard.edu website, but that never works for me either. The download slows to a crawl and never progresses, for some reason.

Any help in how to obtain the database would be much appreciated.

Kind regards,
Imadh

Hi Imadh, Sorry to hear you are having issues with downloading the database from our server. Would you try one more time? Occasionally our server is under a bit of a load which can make downloads take longer. Hopefully when you try again it will all work out okay. If not, please post again.

Thank you,
Lauren

Hi Lauren,

I was able to download the database, but It does seem to have taken quite a few tries. May have been a timing issue, but At least it’s here.

Kind regards,
Imadh

Hi Imadh, Glad to hear you got the database! Sorry it took so many tries. We will see what we can do on our end to make the downloads easier.

Thank you,
Lauren

Hi All, I am facing the issue while downloading the-

humann_databases --download utility_mapping full /usr/software/humann_db --update-config yes
Download URL: http://huttenhower.sph.harvard.edu/humann_data/full_mapping_v201901b.tar.gz
CRITICAL ERROR: Unable to download and extract from URL: http://huttenhower.sph.harvard.edu/humann_data/full_mapping_v201901b.tar.gz

After doing multiple retries since last 1 days finally it is installed, Just want to highlight this issue , I was getting 403 forbidden error for http://huttenhower.sph.harvard.edu/humann_data/full_mapping_v201901b.tar.gz

Hi @Rakesh_Chouhan , Sorry to hear you are having issues with the download. I am glad to hear you finally got it installed. We have been working on migrating our downloads to a new provider. Hopefully they should be online and stable now. Also they should be faster!

Please post if you have any issues!

Thanks!
Lauren

humann v3.7 and MetaPhlAn version 4.0.0 (22 Aug 2022)

Hi, I had downloaded chocophlan and uniref databases October 3rd and they were successful and downloaded in ~40 min. Since then, my databases were deleted by cluster maintenance and I have been unsuccessful in downloading them again for the past 2 days. The command generates the folders for the databases but does not give any output or error message, and does not download any files (currently I had both of them running for 24h).
code:
humann_databases --download chocophlan full $INSTALL_LOCATION
humann_databases --download uniref uniref90_diamond $INSTALL_LOCATION

Is there a problem with the new provider? I appreciate your help, thanks!!

We just recently re-hosted all of our downloads, and there was a period where the downloads were not available. If this issue is still persisting please let us know.

Thank you, I have been trying for the whole week and I wasn’t able to download the databases

With
humann_databases --download chocophlan full $INSTALL_LOCATION
I don’t get an error message or output.
If trying with wget I get this error message:

(humann) $** wget http://huttenhower.sph.harvard.edu/humann_data/chocophlan/full_chocophlan.v201901_v31.tar.gz

–2023-10-30 11:45:56-- http://huttenhower.sph.harvard.edu/humann_data/chocophlan/full_chocophlan.v201901_v31.tar.gz

Resolving huttenhower.sph.harvard.edu (huttenhower.sph.harvard.edu)… 199.94.60.28

Connecting to huttenhower.sph.harvard.edu (huttenhower.sph.harvard.edu)|199.94.60.28|:80… connected.

HTTP request sent, awaiting response… 302 Found

Location: https://huttenhower.sph.harvard.edu/humann_data/chocophlan/full_chocophlan.v201901_v31.tar.gz [following]

–2023-10-30 11:45:56-- https://huttenhower.sph.harvard.edu/humann_data/chocophlan/full_chocophlan.v201901_v31.tar.gz

Connecting to huttenhower.sph.harvard.edu (huttenhower.sph.harvard.edu)|199.94.60.28|:443… connected.

Unable to establish SSL connection.

The admin from Duke Computing Cluster suggests there could be a security block to access your new server. I appreciate if you have any updates on this topic, I am still trying to download the databases (chocophlan and uniref databases). For chocophlan, I am trying with:

conda activate humann
humann_databases --download chocophlan full /hpc/group/humanndatabase

Maybe there is something else we have to specify now? Thank you!

You can use this server instead:
http://cmprod1.cibio.unitn.it/databases/HUMAnN/

I also had the same problem when downloading via a linux terminal from the default server. It works when downloading using a browser though.

Good luck

1 Like

Hi @Emilia , I just wanted to check in to see if you were able to download the databases. If not, it would be great to know what security blocks your admin is seeing. We should not have any on our end but it would be great to know if you are experiencing any issues (or seeing any blocks) so we can get them fixed!

Thanks!
Lauren

If you prefer using the command line, try running the humann_databases command again. Sometimes, transient network issues can cause intermittent failures.
humann_databases --download chocophlan full path/to/directory

I am facing the download error issue. Is there any solution? I hvae been trying it for last few days but getting the same error

1 Like

Syed, today it worked for me: HUMAnN database download - #2 by gjordaopiedade

Thank you @gjordaopiedade I manually downloaded the databases and updated the config file.

now I am facing the next problem i.e., “ERROR: You are using the demo utility database with a non-demo input file. If you have not already done so, please run humann_databases to download the full utility database. If you have downloaded the full database, use the option --utility-database to provide the location. You can also run humann_config to update the default database location. For additional information, please see the HUMAnN User Manual.”

I have all databases (not the demo ones but full databases) download to a directory but I am getting the error on the following commands


“humann3 --input SAMN08516737_1.sam --output …/gene_seqs --nucleotide-database …/…/…/…/…/reference/reference_database/humann3/chocophlan --protein-database …/…/…/…/…/reference/reference_database/humann3/uniref --threads 16”

your help would be highly appreciated.

UPDATE
‘’’
humann3 --input SAMN08516737_1.sam --output …/gene_seqs --nucleotide-database …/…/…/…/…/reference/reference_database/humann3/chocophlan --protein-database …/…/…/…/…/reference/reference_database/humann3/uniref --threads 16 --utility-database …/…/…/…/…/reference/reference_database/humann3
Output files will be written to: /xdisk/syedzaidi/Metagenome_bstPrac/MicrobiomeBestPracticeReview/Metagenomics_analysis/Test_Metagenomic_1/analysis/Gene_based_analysis_onContigs/functional_classification/gene_catalogue/gene_seqs

Process the sam mapping results …

Computing gene families …
Traceback (most recent call last):
File “/home/u13/syedzaidi/anaconda3/envs/MG_analysis2/bin/humann3”, line 33, in
sys.exit(load_entry_point(‘humann==4.0’, ‘console_scripts’, ‘humann3’)())
File “/home/u13/syedzaidi/anaconda3/envs/MG_analysis2/lib/python3.7/site-packages/humann/humann.py”, line 1102, in main
families_file=families.gene_families(alignments,gene_scores,unaligned_reads_count)
File “/home/u13/syedzaidi/anaconda3/envs/MG_analysis2/lib/python3.7/site-packages/humann/quantify/families.py”, line 45, in gene_families
total_all_scores_normalization=alignments.convert_alignments_to_gene_scores(gene_scores,config.count_normalization)
File “/home/u13/syedzaidi/anaconda3/envs/MG_analysis2/lib/python3.7/site-packages/humann/store.py”, line 500, in convert_alignments_to_gene_scores
total_all_scores_normalization=1/total_all_scores*1e6
ZeroDivisionError: division by zero
‘’’

@gjordaopiedade

I am not super sure, but the suggestions of chatGPT are a good place to start:

https://chat.openai.com/share/55e59248-c197-47f3-92f9-c4bfe1777973

Make sure that you have the utility-database downloaded