blue_berry

joined 2 years ago
 

This idea combines a neuro-symbolic AI system (take a LLM and use it to generate logical code; then make inferences from it, see Neural | Symbolic Type) with Attempto Controlled English, which is a controlled natural language that looks like English but is formally defined and as powerful as first order logic.

The main benefit is that the result of the transformation from document/natural language to the logical language would be readable by not IT experts, as well as editable. They could check the result, add their own rules and facts, as well as queries.

I created a small prototype to show in which direction it would be going (heavily work in progress though). What do you think of this? Would love to here your opinions :)

[–] blue_berry@lemmy.world 1 points 2 months ago* (last edited 2 months ago) (1 children)

i’m no longer sure if you’re envisioning a web browser or a website builder. your terminology is all over the place.

I's blurring the line in-between. It's trying to set the interaction with the web on a lower level that is closer to the data. It's like you are live-coding the website you want to use for a specific use-case. But then just call the high-level API-endpoints right away. Basically making the dev-tools and the dev-console of browsers the main way to interact with the web (which assumes a web that is build in a similar fashion).

and no, the semantic web is in no way an an open, global codebase. it’s just a way of structuring html. i know berners-lee wanted the web to be more like what you are describing but the web we have today is not that. you’d need a new protocol.

Yeah, that's true :(

[–] blue_berry@lemmy.world 2 points 2 months ago* (last edited 2 months ago) (3 children)

I don't know. Basically, if you already know what you want, maybe you only want to type down a couple of statements (maybe even from a template or a tutorial that you found online), modify some stuff and then hit enter. And maybe this modifying of language could be the "browsing" part of the browser.

If you look at it like this it would also be immediate and precise. You would only need to add very good code completion tools, e.g. when you click on a noun, you see all the attributes it has in your ontology. Much like in a IDE. There you also "browse" the space of all potential programs with the interface of language with code completion for keywords and defined concepts, which act like links in traditional browsers.

In contrast, the semantic web is like a open, global code base, where everybody can contribute to. And traditional browser could not successfully implement a language interface because the code base had no defined semantic, this would be possible for the semantic web. And using LLMs, it could be propagated into other web paradigms.

[–] blue_berry@lemmy.world 3 points 2 months ago* (last edited 2 months ago) (5 children)

there are already text-based browsers like qutebrowser

hypercard

Awesome! Thanks for the references, didn't know there were already some applications in this direction

 

The fundamental idea of this paper is for ChatGPT-like apps to lose natural language for less energy consumption and more determinism in their answers based on controlled natural languages like ACE; for the user to be able to modify this trade-off-ratio at will based on LLMs (which is not possible when starting from a ChatGPT-like app); and to capture this new paradigm in a new type of browser that has natural language as its primary interface, here called a semantic web-first browser.

 

In its finished state, the software is supposed to work like this:

  • Users start Anthem, which act as a peer in the network
  • Once they have found other peers, they connect to a master node, which coordinates the training
  • All peers then train a model with their uploaded songs
  • Then they generate songs from their local models
  • Each user can now listen to and generate their own AI songs

Would love to hear any feedback from you guys :) Does this make the current AI situation better or worse? I could imagine it acting as a protest against AI companies (similar to what Napster did), but I'm not sure which effect it will have on smaller artists (after all, Napster basically lead to music labels coming under the hood of Apple's ecosystem, which in the end, wasn't so good for artists but on the other hand also brought the Fediverse into being ... I'm not sure).

 

I did a prototype implementation of a “network of ML networks” - an internet-like protocol for federated learning where nodes can discover, join, and migrate between different learning groups based on performance metrics (Repo: https://github.com/bluebbberry/MyceliumNetServer). It's build on Flower AI.

Want do you think of this? It could be used to build a Napster/BitTorrent-like app on this to collaboratively train and share arbitrary machine learning models with other people while keeping data private and only sharing gradients instead of whole models to save bandwidth. Would this be a good counter-weight for big AI companies or actually make things worse?

Would love to hear your opinion ;)

[–] blue_berry@lemmy.world 2 points 5 months ago

Cool. Well, the feedback until now was rather lukewarm. But that's fine, I'm now going more in a P2P-direction. It would be cool to have a way for everybody to participate in the training of big AI models in case HuggingFace enshittifies

[–] blue_berry@lemmy.world 2 points 5 months ago

Yeah thats a good point. Also given that nodes could be fairly far apart from one another, this could become a serious problem.

[–] blue_berry@lemmy.world 1 points 5 months ago

Currently the nodes only recommend music (and are not really good at it tbh). But theoretically, it could be all kinds of machine learning problems (then again, there is the issue with scaling and quality of the training results).

 

Von Neumann’s idea of self-replicating automata describes machines that can reproduce themselves given a blueprint and a suitable environment. I’m exploring a concept that tries to apply this idea to AI in a modern context:

  • AI agents (or “fungus nodes”) that run on federated servers
  • They communicate via ActivityPub (used in Mastodon and the Fediverse)
  • Each node can train models locally, then merge or share models with others
  • Knowledge and behavior are stored in RDF graphs + code (acting like a blueprint)
  • Agents evolve via co-training and mutation, they can switch learning groups and also chose to defederate different parts of the network

This creates something like a digital ecosystem of AI agents, growing across the social web; with nodes being able to freely train their models, which indirectly results in shared models moving across the network in comparison to siloed models of current federated learning.

My question: Is this kind of architecture - blending self-replicating AI agents, federated learning, and social protocols like ActivityPub - feasible and scalable in practice? Or are there fundamental barriers (technical, theoretical, or social) that would limit it?

I started to realize this using an architecture with four micro-services for each node (frontend, backend, knowledge graph using fuseki jena and activitypub-communicator); however, it brings my local laptop to its limits even with 8 nodes.

The question could also be stated differently: how much compute would be necessary to trigger non trivial behaviours that can generate some value to sustain the overall system?

 

cross-posted from: https://lemmy.world/post/28384872

This is a showcase of combining vibe coding with the Fediverse and attempto controlled english (ace).

I'm fascinated by vibe coding, but I'm also highly critical of it. It fascinates me, because it enables people, who normally cannot code to be able to generate running code. What I don't like, is that it just isn't actual programming. It's closer to a wishing well. It fosters a quasi-magical understanding of programming and computer science, which is already too common in current society (I wrote a paper about it here: https://philpapers.org/rec/BINAKR). That's why, in my opinion, the Fediverse should set a counter-point here with something like a first-order logic language like ACE, which actually brings people closer to an actual understanding of computer science concepts like modeling and logic without hiding the complexity behind seemingly "magic", and could also result in better code.

The above demo shows a glimpse of how this could look like on the Fediverse. Imagine communities being able to form their own spaces on the social web through language! Simply using natural language will probably not be specific enough here. We always imagine everything getting much easier, but that's just the logic of digital capitalism that tries to sell us innovation as inventing yet a more easy way to get your coke handed to you, which can only lead to more and more environmental destruction. So, what will the language interface for the future digital look like? I think it will be more something like the semi-formalic language found in technical manuals, cooking recipes and judicial texts. Something like ace, in between coding, domain specific languages, modeling and natural language. And people who are experts at this and know the old technical stuff that no one understands anymore will be the new "coders". But maybe I'm wrong.

Repo: https://github.com/bluebbberry/AceCoding.social.