Ad4mWayn3

joined 2 years ago
 

Let's leave the networking aspect aside for a moment.

When a language is compiled, the source files go through a pipeline of parser -> preprocessor -> compiler -> assembler -> linker, to end up with an executable. With interpreted languages, the source code is instantly executed line by line by an interpreter software. With JIT languages, the program gets compiled and optimized into portable bytecode, which is run by the language's runtime.

If I had to guess, web pages (i.e. HTML/CSS/JS) are most likely run by an interpreter that is a web browser, but isn't that inefficient given that most of what people do on computers is browsing the web? What about browsers, what standard is there that specifies how each language should be run/rendered? What pipeline does a webpage go through to end up as a process in a computer?

 

Is there any computer program with AI capabilities (the generative ones seen in ChatGPT; onlineText-to-Picture generators, etc.) that is actually standalone? i.e. able to run in a fully offline environment.

As far as I understand, the most popular AI technology rn consists of a bunch of matrix algebra, convolutions and parallel processing of many low-precision floating-point numbers, which works because statistics and absurdly huge datasets. So if any such program existed, how would it even have a reasonable storage size if it needs the dataset?