Chief Delphi

Chief Delphi (http://www.chiefdelphi.com/forums/index.php)
-   Programming (http://www.chiefdelphi.com/forums/forumdisplay.php?f=51)
-   -   New compression method (http://www.chiefdelphi.com/forums/showthread.php?t=30187)

Aalfabob 06-09-2004 23:04

New compression method
 
I made a new way to compress files a couple of days ago, ive made the program and have tested them. The only problem is that most people will not believe that what i have done is possible and i cant actually send them how it works because i havent patented it yet. But heres my question, If you are able to compress the same file over and over again would it be possible to reach very small sizes? Ive tryed a 1 Mb file so far and it reached 515 bytes, and it is fully reversable.

Max Lobovsky 06-09-2004 23:20

Re: New compression method
 
It's possible that recompressing a file can make it smaller, depending on the compression scheme used, but, <edit>I think</edit> most modern compression schemes compress as much as they can on there first pass (sometimes because there first pass really includes several passes at compressing the data).

Anyway, compressing a file from 1Mb to 515 bytes doesn't really say anything about your compression scheme. If you give me a file of any size, I can very simply write a compression scheme to compress it to 0 bytes. If you can take arbitrary files and consistently compress them to a small size, then you have a compression scheme of merit.

ahecht 07-09-2004 00:34

Re: New compression method
 
Quote:

Originally Posted by Aalfabob
I made a new way to compress files a couple of days ago, ive made the program and have tested them. The only problem is that most people will not believe that what i have done is possible and i cant actually send them how it works because i havent patented it yet. But heres my question, If you are able to compress the same file over and over again would it be possible to reach very small sizes? Ive tryed a 1 Mb file so far and it reached 515 bytes, and it is fully reversable.

I don't know. When a brand new member with no team number comes in here and, in their very first post, posts the equivalent of digital snake oil, I become a bit suspicious. I'm not sure what this guy is trying to pull (such as whether he is an existing member who set up a dummy account for this or just a random troll), but I can assure people that something is up.

Besides, with most compression methods, compressing an already compressed file results in a slightly larger file size.

Ryan M. 07-09-2004 07:08

Re: New compression method
 
Quote:

Originally Posted by ahecht
I'm not sure what this guy is trying to pull (such as whether he is an existing member who set up a dummy account for this or just a random troll), but I can assure people that something is up.

I agree that it is a little unbelievable. Alaphabob, I'd trust almost anyone here with something that isn't protected yet. Just ask anyone with 5+ "rep(utation) points" to look over your algorithm. (Those are the green dot's on the left of the darker-grey bar above their post.) I'm not saying that less rep'd people aren't trustworthy, but it's something to reasure you.
Quote:

Originally Posted by ahecht
Besides, with most compression methods, compressing an already compressed file results in a slightly larger file size.

That's because, with all most all compression schemes, they work by creating a library, at some point in the file, and put multi-byte strings into it. The rest of the file is then encoded by putting a one byte key in place of the original string. If this string happens more than once, then you save space. For instance, if you have a string in a text file like "happy" three times, then you put "happy" in your library once and 3 one byte markers in the text for a total of around 8 bytes (there's probably also buts separating different items in the library, etc, which is why I say "around"). The original three happies took up 15 bytes.

When you recompress the file, it ends up compressing a file with no, or very few, redundencies, which are what make the library method work so well.

EDIT:
Why the heck did I choose happy? Why not something cool, like FIRST? :D

Aalfabob 07-09-2004 15:37

Re: New compression method
 
Alright, first off I'm not trying to pull anything. Why would I create an account and waste my time to mess around with people, a friend gave me this site to ask a few people if they think it would be possible. Second off, my compression theorm isnt like the others, it doesnt run on how many times certain charactors show up in a file as a hole. This makes it able to recompress the same file over and over again almost always gaining a compression. This also means that it can work on any type of file, .zip, .exe, .jpg, ect. But it does reach a limit to the file sizes it can reach, with the current program I have made it can compress any file type and size down to 508 bytes and ussually fluctuates around 508 and 515 bytes. Because a file is larger then another doesnt mean it can not hit this limit, it just means that more attemps must be made to reach it. I have some data charts if anyone wishes to see them.

Jeff_Rice 07-09-2004 15:45

Re: New compression method
 
Is it a lossy or lossless?

Aalfabob 07-09-2004 15:48

Re: New compression method
 
Lossless, what would be the point of compressing a data file if it was corrupted when uncompressed?

I am going to see if any large buisnesses are interested in this and if not I will make it open source, this is one of the reasons why I am trusting noone. Even if someone is trustful, there is still always that very small chance of it getting out.

Jeff_Rice 07-09-2004 16:06

Re: New compression method
 
Well, why did you post anything at all if you are trusting no one? Without your algorithm it is very difficult to help you.

Compression is like factoring. In factoring, you take a complex equation and can define it simply by its solutions. In compression you do a similar thing. However, you will eventually run into a floor no matter how good a compression system you use. This is due to the fact that the information is still there, just in a compressed format. I am guesing your algorithm has some sort of system for recording what it has done in order that it can be undone. This file created requires space. The more times you compress, the closer the file is to becoming "prime" toward the algorithm. Eventually you reach a point where the information to expand the file makes the file large enough that compression will not make the whole set any smaller.

So basically what ryan morehart said, but in more generic terms.

Aalfabob 07-09-2004 16:36

Re: New compression method
 
I understand the problem of you not being able to help me because im not releasing how it works. Are there any ways that I would be able to make it open source for a little while till I am able to open it up for commericial uses? I want to be able to keep the theorm if I ever decide to make money with it and I want to make sure noone can steal it. Would making it open source save it from being stolen? Ive looked into patents but there is no way I can afford the $4000 to get one and the $1500 to keep them updated every couple of years. If anyone has a link or something to help me out here please post it.

Ill be happy to post all the information needed about it as soon as its safe. And I do understand that this seems impossible but trust me its not :) .

Ryan M. 07-09-2004 16:44

Re: New compression method
 
Well, here's a site which lists the most commonly used open source licenses. Read through them and see what you like. Make sure you choose one which prevents the commercial reuse of the source code.

Edit:
Hm, actually, according to them, "open source" licenses do not prevent commercial use. Whatever... :rolleyes:

seanwitte 07-09-2004 16:48

Re: New compression method
 
Go to www.maximumcompression.com and run your utility against their test files. You'll be able to compare your results against a fairly large set of benchmarks. Post your results. If you really beat those benchmarks then you'll need to have a few volunteers verify your results. For that you can distribute a binary without source and a non-disclosure.

Aalfabob 07-09-2004 16:56

Re: New compression method
 
Alright let me rebuild my program (7 days max) because i have a couple new ideas i would like to try out with it. And i will post up the scores by then or earlyer. Hopefully I will be able to get it done alot sooner but it depends on how much work I have.

Joe Ross 07-09-2004 18:42

Re: New compression method
 
Quote:

Originally Posted by seanwitte
Go to www.maximumcompression.com and run your utility against their test files. You'll be able to compare your results against a fairly large set of benchmarks. Post your results. If you really beat those benchmarks then you'll need to have a few volunteers verify your results. For that you can distribute a binary without source and a non-disclosure.

I'm suprised that site doesn't have a file full of pseudo-random data. While very complete in testing different programs, the files it chooses seem rather arbitrary.

Max Lobovsky 07-09-2004 19:08

Re: New compression method
 
Quote:

Originally Posted by Joe Ross
I'm suprised that site doesn't have a file full of pseudo-random data. While very complete in testing different programs, the files it chooses seem rather arbitrary.

I'm not sure what you mean by "pseudo", but it's mathematically impossible for a compressor to consistently compress random data. I don't know much information theory at all, but i know that random data is essentially pure information, and there is no way to encode pure information into a smaller amount of information. (This is incidentally why already compressed files (including lossy media compression for images, sound, etc) don't compress well. They have already converted the data to near pure information.)

The site's choice of files seems pretty logical to me. It has a selection of files that are commonly compressed. It might be interesting if they tried recompressing some very common compressed file(s), maybe like the Fedora Linux distribution, or Microsoft Office 2k3.

Aalfabob 07-09-2004 19:28

Re: New compression method
 
Quote:

Originally Posted by Max Lobovsky
I'm not sure what you mean by "pseudo", but it's mathematically impossible for a compressor to consistently compress random data. I don't know much information theory at all, but i know that random data is essentially pure information, and there is no way to encode pure information into a smaller amount of information. (This is incidentally why already compressed files (including lossy media compression for images, sound, etc) don't compress well. They have already converted the data to near pure information.)

Well see :) , sometimes the impossible can be done.


All times are GMT -5. The time now is 09:38.

Powered by vBulletin® Version 3.6.4
Copyright ©2000 - 2017, Jelsoft Enterprises Ltd.
Copyright © Chief Delphi