One of the caveats for PAQ and CMIX is that they take huge amounts of RAM and CPU power to compress a little better than LZMA / XZ / 7zip, BWT, RAR, and the long list of all those LZ variant algorithms that are out there.
But what if those resources were safely allocated away from the computer with a dedicated SBC (single board computer) and implemented as an EEPROM or an FPGA either as PCIe or a USB 2.0+ device?
If a device preallocates the storage needed for a file passed to it as a temp file over an SLC SSD and then uses DDR4 RAM up to 32GB on board or higher to meet the RAM requirements for CMIX and PAQ currently...would it be fast enough to use this as a hardware circuit running in parallel to computer systems using <1% of the system CPU to watch it?
(Basically, one can look at this as a hardware implementation of telling a computer with resources on a network: "I want you to grab this file, make a copy of it from the host to your local machine. Use all the ram and cpu that you have there as-needed, but keep me updated on estimated completion time and then pass the compressed file back over the network to me when you're done with it. Don't use any of the resources on the host machine and contain all that over there, but when it's ready send the output here to the host over the network")
The network would be, in this case, the hardware bus for any device or implementation that does this, and the remote machine would be the dedicated SBC.
As a hardware circuit it wouldn't need any resources from the host machine, and could work even on older 386 or 486 systems with a usb card...but would still work hundreds to thousands of times faster per cycle than it does with today's best laptops and desktops when running programs as close to bare metal as possible.
It would likely make things like LZ4 and LZTurbo seem near-instant, but the time-consuming and (sometimes) out of reach resources becomes attainable and modular to any device with a usb connection when a dedicated computer can be added instantly to make those hours, days, or weeks (!) running some algorithms only take seconds, minutes, or days instead.
The only thing that would run on the host would be a type of driver manager program to monitor things, queue compression requests, and move compressed files to the destination directory after they're completed.
Anyone without the device but with adequate hardware (32GB or greater if required for RAM) could still compress and decompress more slowly of course, but having a device like this would make long compression cycles far more productive and easier to test results from, and help streamline compression development if there were a dedicated device for development.
BUT is this still practical in today's world to do this?
Or have "tablets" and "mobile devices" as they have been marketed today to the masses ruined the ability to implement this even as a usb device mainstream after the 2010's?