<?xml version="1.0" encoding="utf-8"?>
<feed xml:lang="en" xmlns="http://www.w3.org/2005/Atom"><title>Recent changes to support-requests</title><link href="https://sourceforge.net/p/prinseq/support-requests/" rel="alternate"/><link href="https://sourceforge.net/p/prinseq/support-requests/feed.atom" rel="self"/><id>https://sourceforge.net/p/prinseq/support-requests/</id><updated>2012-01-20T20:08:49Z</updated><subtitle>Recent changes to support-requests</subtitle><entry><title>Turn off duplicate removal</title><link href="https://sourceforge.net/p/prinseq/support-requests/4/" rel="alternate"/><published>2012-01-20T20:08:49Z</published><updated>2012-01-20T20:08:49Z</updated><author><name>Anonymous</name><uri>https://sourceforge.net/u/userid-None/</uri></author><id>https://sourceforge.net8a193c1e95d578da4f7d5621e087d442baa5b7bf</id><summary type="html">&lt;div class="markdown_content"&gt;&lt;p&gt;I am working with illumina metatranscriptomic fastq files at around 8 GB. The vast majority of my exact duplicates are ribosomal sequences that I would like to keep in my quality filtered dataset with prinseq. Is there a way to turn off the duplicate removal in the standalone version?&lt;/p&gt;&lt;/div&gt;</summary></entry><entry><title>processing after analysis on webversion</title><link href="https://sourceforge.net/p/prinseq/support-requests/3/" rel="alternate"/><published>2011-10-18T07:06:10Z</published><updated>2011-10-18T07:06:10Z</updated><author><name>Anonymous</name><uri>https://sourceforge.net/u/userid-None/</uri></author><id>https://sourceforge.net39a018969bc43890a78b83670cf5aae653837c52</id><summary type="html">&lt;div class="markdown_content"&gt;&lt;p&gt;I really really like your site. I've a rather silly question though. After uploading a fasta file I get all the stats but I don't understand how to run the file with the filter and trim options. So I'd like to analyze a fasta file and then do some seperate sessions of trimming.&lt;br /&gt;
best regards and thanks foor your support!&lt;/p&gt;&lt;/div&gt;</summary></entry><entry><title>problems with large files</title><link href="https://sourceforge.net/p/prinseq/support-requests/1/" rel="alternate"/><published>2011-10-13T13:21:48Z</published><updated>2011-10-13T13:21:48Z</updated><author><name>Javier Alonso</name><uri>https://sourceforge.net/u/fjalonso1234/</uri></author><id>https://sourceforge.netcf8a0cc806ebd44c567f94e6076f66a5e0897153</id><summary type="html">&lt;div class="markdown_content"&gt;&lt;p&gt;We are using prinseq-lite to filter fastq files with the parameters -derep 1. The function works fine with files less that 3 GB, but I have problems when filtering largest files (I think that the computer is probably out of memory). I have a computer with a i7 intel processor and 16 GB RAM. The host OS is Windows 7 and I run prinseq on a Virtual Machine (VMware) loaded with Centos 5. I have assigned 8 GB RAM to the Virtual Machine. Could it be a problem of memory?&lt;br /&gt;
Thank you very much in advance&lt;/p&gt;&lt;/div&gt;</summary></entry></feed>