dazzdb 1.0-1 source package in Ubuntu

Changelog

dazzdb (1.0-1) unstable; urgency=low

  * Initial release (Closes: #798873)

 -- Afif Elghraoui <email address hidden>  Sun, 20 Sep 2015 03:11:58 -0700

Upload details

Uploaded by:
Debian Med on 2015-10-16
Uploaded to:
Sid
Original maintainer:
Debian Med
Architectures:
any
Section:
misc
Urgency:
Low Urgency

See full publishing history Publishing

Series Pocket Published Component Section
Xenial release on 2015-10-24 universe misc

Downloads

File Size SHA-256 Checksum
dazzdb_1.0-1.dsc 1.8 KiB 943527e3421ea21c83c57dbe4cff75093c0679b59dcef7b0271fec8512c3240f
dazzdb_1.0.orig.tar.gz 61.7 KiB ae7a71b71925dd5a1fd4026ca7462a70cb8ade9d3afd5cdeef0f1bd11b9c9105
dazzdb_1.0-1.debian.tar.xz 8.1 KiB 24f265f7a024949006f75ea8b9f7e139fafea192503a93c0ccbd613b77c3123e

No changes file available.

Binary packages built by this source

dazzdb: manage nucleotide sequencing read data

 To facilitate the multiple phases of the dazzler assembler, all the read data
 is organized into what is effectively a database of the
 reads and their meta-information. The design goals for this data base
 are as follows:
  * The database stores the source Pacbio read information in such a
    way that it can re-create the original input data, thus permitting
    a user to remove the (effectively redundant) source files. This
    avoids duplicating the same data, once in the source file and once
    in the database.
  * The data base can be built up incrementally, that is new sequence
    data can be added to the data base over time.
  * The data base flexibly allows one to store any meta-data desired for
    reads. This is accomplished with the concept of *tracks* that
    implementors can add as they need them.
  * The data is held in a compressed form equivalent to the .dexta and
    .dexqv files of the data extraction module. Both the .fasta and
    .quiva information for each read is held in the data base and can be
    recreated from it. The .quiva information can be added separately and
    later on if desired.
  * To facilitate job parallel, cluster operation of the phases of the
    assembler, the database has a concept of a *current partitioning* in
    which all the reads that are over a given length and optionally
    unique to a well, are divided up into *blocks* containing roughly a
    given number of bases, except possibly the last block which may have
    a short count. Often programs can be run on blocks or pairs of blocks
    and each such job is reasonably well balanced as the blocks are all
    the same size. One must be careful about changing the partition
    during an assembly as doing so can void the structural validity of
    any interim block-based results.

dazzdb-dbgsym: debug symbols for package dazzdb

 To facilitate the multiple phases of the dazzler assembler, all the read data
 is organized into what is effectively a database of the
 reads and their meta-information. The design goals for this data base
 are as follows:
  * The database stores the source Pacbio read information in such a
    way that it can re-create the original input data, thus permitting
    a user to remove the (effectively redundant) source files. This
    avoids duplicating the same data, once in the source file and once
    in the database.
  * The data base can be built up incrementally, that is new sequence
    data can be added to the data base over time.
  * The data base flexibly allows one to store any meta-data desired for
    reads. This is accomplished with the concept of *tracks* that
    implementors can add as they need them.
  * The data is held in a compressed form equivalent to the .dexta and
    .dexqv files of the data extraction module. Both the .fasta and
    .quiva information for each read is held in the data base and can be
    recreated from it. The .quiva information can be added separately and
    later on if desired.
  * To facilitate job parallel, cluster operation of the phases of the
    assembler, the database has a concept of a *current partitioning* in
    which all the reads that are over a given length and optionally
    unique to a well, are divided up into *blocks* containing roughly a
    given number of bases, except possibly the last block which may have
    a short count. Often programs can be run on blocks or pairs of blocks
    and each such job is reasonably well balanced as the blocks are all
    the same size. One must be careful about changing the partition
    during an assembly as doing so can void the structural validity of
    any interim block-based results.