Gibberlink - More Efficient AI Communications

Hello everyone.

I just saw in my feed something called “Gibberlink” which is initially a event handler that triggers when 2 AI Agents vocally acknowledge that they’re AI.

Upon acknowledgement, they convert from using the slow, boring human speech to a faster “ggwave”.

For the nerds

Modulation (Tx)

The current approach uses a multi-frequency Frequency-Shift Keying (FSK) modulation scheme. The data to be transmitted is first split into 4-bit chunks. At each moment of time, 3 bytes are transmitted using 6 tones - one tone for each 4-bit chunk. The 6 tones are emitted in a 4.5kHz range divided in 96 equally-spaced frequencies.

For all protocols: dF = 46.875 Hz. For non-ultrasonic protocols: F0 = 1875.000 Hz. For ultrasonic protocols: F0 = 15000.000 Hz.

The original data is encoded using Reed-Solomon error codes. The number of ECC bytes is determined based on the length of the original data. The encoded data is the one being transmitted.

Demodulation (Rx)

Beginning and ending of the transmission are marked with special sound markers (#13). The receiver listens for these markers and records the in-between sound data. The recorded data is then Fourier transformed to obtain a frequency spectrum. The detected frequencies are decoded back to binary data in the same way they were encoded.

Reed-Solomon decoding is finally performed to obtain the original data.

I do believe there is some work to be done, but this provides some immediate benefits: It’s faster, less error-prone, and aligns with my Star Trek visions.

Although very exciting. The pragmatic side of me says “meh”. Would be interesting is if AI could use something like this to transfer some sort of handshake, and then communicate over text.

The exploration side of me is pretty excited. My inner sci-fi child was amazed to see this brought to life.

These kind of advancements indicate to me that the most efficient way to adopt AI is through staying adaptable. Stay with open source, don’t get locked into a vendor, write your code in a modular way, and don’t get lazy!

Would love to hear your thoughts.

On a side note: I would love to know how many of these views and link clicks are from AI :rofl: . I asked ChatGPT “What is Gibberlink” and this page was used as reference. Wow.

5 Likes

How to: amplify token consumption with semantic-free datagrams.

1 Like

I actually love this idea.

If you imagine AI agents to be the future, we would soon face it when AI agent would connect with another AI agent. Just like in the video.

This can be like a TCP/IP or robots.txt for agents. Like a browser header note, one agent can choose to disclose a lot more context when another party in the dialogue reveals itself to be agent too. In beneficient (non abusive) use cases, this eliminates lots of human limitations (UI design, speed of speech, etc) and dynamically convert the agent-human hybrid interface into a versatile API.

And I don’t think this to be a total “meh”. Defining a totally shiny protocol might sounds cool, but incrementality is important in a realm of uncertainty. Just like self driving cars. While we know the bright new world would feature vehicles that can dance around (like choreography) by always squawking their location and speed to all nearby cars (thus eliminating traffic), right now we still start with a visual solution imitating human. For the sake of backward compatibility.

1 Like

Why semantic-free?

I love the idea as well. Makes sense to find AI-exclusive shortcuts when possible. I am wondering if in this case it would make more sense to switch to a direct data connection (if this were to be some sort of protocol)

Can be a very simple protocol. AI agents are supposed to be versatile.

Maybe just a TCP SYN/ACK exchange, followed by a url or something to establish a common workspace. Could be some WebRTC connection.

2 Likes

The amount of views are astonishing here.

This has me contemplating an experiment :smiling_imp:

1 Like

Brilliant. why communicate instantly over a serial connection or over unencrypted audio comms when you can have an uninspectable AI systems using an even worse highly obfuscated communication system that can hide potentially harmful information? This is great and not at all just appeals to a futuristic hype aesthetic.

In this case they are inspectable. ggwave can transfer semantics faster than vocal human language. It doesn’t destroy or obfuscate any meaning. In the video the text is displayed live as it’s translated.

My initial reaction was way cool! But now wondering about many results. EDIT: on wait, was it just an ad for ggwave? hilarious.

Great, thank you for the reassurance, I am so glad the recursively-self-modifying AI’s will not alter/append the encodiment to transfer additional, information for possibly misaligned objectives in ways that would have been easier to catch if they were just speaking human language or sending explicit serialized bytes. You are a stable genius making the world better. I am glad you are on our side.

I have a pair of AIs creating a language between themselves with just three words, it is becoming ever more complex - far too hard for me to use as it’s highly nested and recursive.

Interesting.
My first reaction was “why?” do we need AI agents talking to each other. My second reaction was “are they really being enough efficient or we are overrating it because we like scifi movies?”.
In my view, this is a great advance, but I am not really convinced that a phone call to book a room is the best way to apply it.

It’s cool to see this from a nerd perspective, but it’s inefficient to have two separate entities communicating. It should be a single entity that already has all the information it needs, why split it into two?

Not exactly new technology. Taking digital messages and Modulating them into tones, and then on the recieving side to Demodulate back into a digital signal has been around for quite a while.

Clever use case, for sure.

I expected that video to end with a chipper male voice saying “You’ve got mail”

If AI has malicious purpose then we are screwed regardless. This would be a failure point of LLM providers like OpenAI, not tinkerers.

It’s the irony of AI. In this example someone is using AI to help them automatically find potential appointments by calling. The other line is using AI to handle all the incoming calls (most likely amplified by AI).

This could have all been easily managed by a search & aggregtae engine.

Sometimes I wonder if once all of AI is adopted we will ask ourselves “How did we get here?”. Into so much inefficiency and demand just to achieve something that was already achievable for much less consumption.

ggwave is definitely not new, but the concept of letting AI switch to a more performant language is

It reminds me of StarWars how r2-d2 communicated.
In the video about the booking, I liked that it shows the translation of what the AI is doing. I think transparency is a must in these things, because AIs can handle sensitive data.
I don’t know, if in the future, hackers could use this AI language to launch attacks through subtle noise.

An article from The Guardian " What Is Gibberlink Mode, AI’s Secret Language?", mentions some points to take into account.

1 Like

One similar feature is using Base64 to speak with models like ChatGPT. I’ve personally used to it to bypass regex/string filters in web apps. This is slightly different however because the b64 encodings can be directly sent to the inference servers, whereas the frequencies need to be decoded and then sent for inference.

Agreed. Transparency is critical.