Ocarina Networks Patent applications |
Patent application number | Title | Published |
20110125722 | METHODS AND APPARATUS FOR EFFICIENT COMPRESSION AND DEDUPLICATION - Mechanisms are provided for performing efficient compression and deduplication of data segments. Compression algorithms are learning algorithms that perform better when data segments are large. Deduplication algorithms, however, perform better when data segments are small, as more duplicate small segments are likely to exist. As an optimizer is processing and storing data segments, the optimizer applies the same compression context to compress multiple individual deduplicated data segments as though they are one segment. By compressing deduplicated data segments together within the same context, data reduction can be improved for both deduplication and compression. Mechanisms are applied to compensate for possible performance degradation. | 05-26-2011 |
20110125719 | EFFICIENT SEGMENT DETECTION FOR DEDUPLICATION - Mechanisms are provided for efficiently detecting segments for deduplication. Data is analyzed to determine file types and file components. File types such as images may have optimal data segment boundaries set at the file boundaries. Other file types such as container files are delayered to extract objects to set optimal data segment boundaries based on file type or based on the boundaries of the individual objects. Storage of unnecessary information is minimized in a deduplication dictionary while allowing for effective deduplication. | 05-26-2011 |
20110082840 | SCALABLE MECHANISM FOR DETECTION OF COMMONALITY IN A DEDUPLICATED DATA SET - Mechanisms are provided for efficiently determining commonality in a deduplicated data set in a scalable manner regardless of the number of deduplicated files or the number of stored segments. Information is generated and maintained during deduplication to allow scalable and efficient determination of data segments shared in a particular file, other files sharing data segments included in a particular file, the number of files sharing a data segment, etc. Data need not be expanded or uncompressed. Deduplication processing can be validated and verified during commonality detection. | 04-07-2011 |
20100094813 | REPRESENTING AND STORING AN OPTIMIZED FILE SYSTEM USING A SYSTEM OF SYMLINKS, HARDLINKS AND FILE ARCHIVES - A data de-duplication system is used with network attached storage and serves to reduce data duplication and file storage costs. Techniques utilizing both symlinks and hardlinks ensure efficient deletion file/data cleanup and avoid data loss in the event of crashes. | 04-15-2010 |
20100088277 | OBJECT DEDUPLICATION AND APPLICATION AWARE SNAPSHOTS - Embodiments deploy delayering techniques, and the relationships between successive versions of a rich-media file become apparent. With this, modified rich-media files suddenly present far smaller storage overhead as compared to traditional application-unaware snapshot and versioning implementations. Optimized file data is stored in suitcases. As a file is versioned, each new version of the file is placed in the same suitcase as the previous version, allowing embodiments to employ correlation techniques to enhance optimization savings. | 04-08-2010 |
20090289819 | EFFICIENT DATA COMPRESSION & DECOMPRESSION OF NUMERIC SEQUENCES - Embodiments described herein relate to compression and decompression of data consisting of a one dimensional time series of floating point numbers. A compressor may comprise a lossless stage and in some embodiments a lossy stage in addition to the lossless stage. The lossy stage quantizes the data by discarding some of the least significant bits as specified by the user. The lossless stage uses a context mixing algorithm with two bit-wise predictive models whose predictions are combined and fed to an arithmetic coder. One model is a direct context model using the most significant bits of prior numeric samples as context. The other model is the output of an adaptive filter, in which the approximate predicted numeric value is used as context to model the actual value. A corresponding decompressor uses the same lossless model with the arithmetic coder replaced by an arithmetic decoder. | 11-26-2009 |
20090240718 | DEFLATE FILE DATA OPTIMIZATION - Mechanisms are provided for optimizing file data compressed using deflate mechanisms such as the ZLIB Compressed Data Format Specification and the DEFLATE Compressed Data Format Specification. Deflate mechanisms output different deflate file data depending on specific file data parameters. An optimization tool decompresses the deflate file data and outputs an optimized data stream. When a client application attempts to access the deflate data, the tool deoptimizes the optimized data stream and applies the same deflate algorithm to generate deflate file data. Although the deflate algorithm is applied without using the file data parameters used to generate the original deflate file data, substitute deflate file data is produced. | 09-24-2009 |
20090216788 | MULTIPLE FILE COMPACTION FOR NETWORK ATTACHED STORAGE - Mechanisms are provided for optimizing multiple files in an efficient format that allows maintenance of the original namespace. Multiple files and associated metadata are written to a suitcase file. The suitcase file includes index information for accessing compressed data associated with compacted files. A hardlink to the suitcase file includes an index number used to access the appropriate index information. A simulated link to a particular file maintains the name of the particular file prior to compaction. | 08-27-2009 |
20090216774 | VIRTUALIZATION OF METADATA FOR FILE OPTIMIZATION - Mechanisms are provided for optimizing files while allowing application servers access to metadata associated with preoptimized versions of the files. During file optimization involving compression and/or compaction, file metadata changes. In order to allow file optimization in a manner transparent to application servers, the metadata associated with preoptimized versions of the files is maintained in a metadata database as well as in an optimized version of the files themselves. | 08-27-2009 |