Image and video compression has become a field that is painfully hard to enter.
State of the art is complex and exhaustive, the functionality of reference encoders and comments/versions among them is really a lot.
We are well beyond where a dedicated individual can try an idea, show that it is better and expect that others can pick it up (e.g. in standardization). It is not sufficient to run a few dozen images and judge by yourself, you are expected to demonstrate the benefit integrated into the latest reference encoders and need a sponsor to join standardization efforts.
For educational purpose? Sure - do whatever you want - but any discussion "is it novel" or "is it useful for others" is moot, unfortunately.
Yes, but many of those complex state of the art thing are comprised of parts that began as ideas like this. All ideas have to start somewhere.
You don't have to have a idea that beats everything in all circumstances. It doesn't even have to be better on even one thing to potentially be useful to others. Reading this article gave me some ideas related to a supplementary autoencoder I had been working with, where encode is ( A + B ) -> Latent, decode is (A+Latent) -> B . Which is essentially the same problem If you consider the A to be the Luma, B to be the Chroma, and the compressed Chroma to be the Latent.
In that respect, I found it useful to see a perspective on an analogous problem. That makes the "Is it useful for others" not only answerable, but true.
I applaud the effort of tinkering, re-creating and sharing, but I think the name is misleading - it is not at all a "local GPT". The contribution is not to do anything local and it is not a GPT model.
Oh... Having a local-only voice assistant would be great. Maybe someone can share the practical side of this.
Do you have the GPU running all day at 200W to scan for wake words? Or is that running on the machine you are working on anyway?
Is this running from a headset microphone (while sitting at the desk?) or more like a USB speakerphone? Is there an Alexa jailbreak / alternative firmware as a frontend and run this on a GPU hidden away?
It is a new neat idea to selectively adjust focus distance for different regions of the scene!
- processing: while there is no post processing, it needs scene depth information which requires pre computation, segmentation and depth estimation. Not a one-shot technique and quality depends on computational depth estimates being good
- no free lunch. The optical setup needs to trade in some light for this cool effect to work. Apart from the limitations of the prototype, how much loss is expected in theory?
How does this compare to a regular camera setup with lower aperture? F/36 seems excessive for comparison.
- resolution - what resolutions have been achieved? (maybe not the 12 MPixels of the sensor? For practical or theoretical reasons? ) What depth range can the prototype capture? "photo of Paris Arc de triumphe displayed on a screen". This is suspiciously omitted
- how does the bokeh look like when out of focus? At the edge of an object?
The introduction of weird or unnatural artifacts would seriously limit the acceptance
Don't get me wrong - nice technique! But to my liking the paper is omitting fundamental properties
While this methods has no post processing, it requires a pre processing step to pre-calture the scene, segment it, estimate depth an compute the depth map.
You try to compute something that supercomputer can't - by not computing it? Instead the formula is stored in a data structure.
But once you need to access all the values you still have something that does not fit the memory and needs to be computed.
I can't judge on the Java side, but suggest to pick a better example on how this can be useful.
Most languages force you to choose: either compute everything upfront (O(n) memory) or write complex lazy evaluation code. Coderive gives you declarative intent with automatic optimization. You write what you mean (for i in [0 to 1Qi]), and the runtime figures out the optimal execution strategy. This is like having a JIT compiler that understands mathematical patterns, not just bytecode."
it only compute what was needed on the right time. See this output for example:
Enter file path or press Enter for default [/storage/emulated/0/JavaNIDE/Programming-Language/Coderive/executables/LazyLoop.cod]:
>
Using default file: /storage/emulated/0/JavaNIDE/Programming-Language/Coderive/executables/LazyLoop.cod
Testing timer() function:
Timer resolution: 0.023616 ms
Linux gpib also supports my adapter natively with a special linux gpib firmware you can download from the page above. It also supports multiple instruments connected to one adapter then.
This sounds more like a Matt Parker video idea - get a bunch of people, three theodolites to measure angles accurately, a good location and start measuring angles for line of sight and see how well this determines the earth's radius.
Rough estimate - with an excellent 0.5" angular resolution and 35km triangle this could work.
It would be interesting to learn how this was created.
Did you buy all these colors and paint and scan them?
Did you analyze the shopping images of the bottles and classify them into hex colors?
Or maybe just group by the color names given in the storefront listing?
Vastly different efforts, different "accuracy", but still, each methods has its use. But knowing what to expect would be nice.
I pulled out the RGB values from the solid color areas of the swatch images on the manufacturers' websites for each of the colors.
It's definitely just an approximation of the real-world color, but I figured that if that's the RGB value the manufacturer used, it's probably pretty close.
Then I calculate the euclidean distance between the RGB value from the provided hex and each of the paints, and show the two closest matches from each brand.
Closed circuit television (CCTV) is a term to describe video transmission that is not broadcast. Traditionally with BNC cables going to a control room, monitors and recorders.
I think this software-only post is meant for IP cameras / surveillance cameras. Internet is the oposite of closed circuit.
Maybe CCTV is used as a synonym for surveillance now in some regions of the world, but certainly confusing for a non-native speaker.
> I think this software-only post is meant for IP cameras / surveillance cameras. Internet is the oposite of closed circuit.
I think in this case, IP is referring to IP from TCP/IP, meaning "The Internet Protocol", not necessarily over/through "public internet links", so as long as you're only within your own local network/WAN, wouldn't that still be CCTV then? Or maybe the "closed circuit" thing is more of a physical property than I read it to be?
We are well beyond where a dedicated individual can try an idea, show that it is better and expect that others can pick it up (e.g. in standardization). It is not sufficient to run a few dozen images and judge by yourself, you are expected to demonstrate the benefit integrated into the latest reference encoders and need a sponsor to join standardization efforts.
For educational purpose? Sure - do whatever you want - but any discussion "is it novel" or "is it useful for others" is moot, unfortunately.
reply