Technology has long influenced the genesis of language and culture, which traces back to the earliest types of writing.

The medium itself – cave partitions, stone tablets, or paper – shaped how language was used and perceived.

Today, AI and its related terms are entering the lexicon, underscoring its rising cultural impact. Generative AI like ChatGPT, once the mutterings of a select few AI enthusiasts, has quickly change into a household name accessed by billions of monthly users. 

Reflecting the impact of AI on popular culture, the Cambridge Dictionary recently named “hallucinate” as their word of the yr, adding a brand new AI-centric definition; “When a synthetic intelligence (= a pc system that has a number of the qualities that the human brain has, comparable to the flexibility to provide language in a way that seems human) hallucinates, it produces false information.” 

Merriam-Webster followed suit, naming “authentic” as its word of the yr, describing the way it’s change into increasingly difficult to find out whether information or content is real or not, partly attributable to the influence of AI deep fakes. 

AI technology also parallels influential technologies from the past and their cultural-linguistic impacts. For instance, the arrival of the printing press within the fifteenth century revolutionized language, introducing recent concepts like typography, punctuation, and standardized spelling.

As technology evolved, so did language, adapting to every recent medium’s constraints and possibilities. In the twenty first century, the web and digital communication have created quite a few neologisms and mixing of words, comparable to the widespread use of prefixes like “cyber-” and “e” as in eCommerce and email.

Exploring AI’s impact on linguistics presents a surface-level view of how technology influences language. 

But if you happen to delve deeper, you realize this linguistic shift is just the tip of the iceberg.

It raises a broader commentary on how AI influences language and culture now, and might change into the driving force behind cultural genesis and knowledge creation in the longer term. 

AI and cultural and knowledge genesis 

In an essay published by The Economist, the authors compare AI’s internalization of information and culture to the labyrinthine corridors of Argentinean creator and librarian Jorge Luis Borges’ “The Library of Babel,” an infinite expanse of hexagonal rooms holding the vastness of human potential and folly. 

The Library of Babel – generated with MidJourney

Each room, lined with books stuffed with every conceivable arrangement of letters and symbols, represents the limitless permutations of each knowledge and nonsense. This library, a metaphor for the universe, is an allegory of each the pursuit of meaning and the overwhelming abundance of data.

A parallel unfolds on the planet of AI, particularly within the sprawling networks of generative AI. 

Frontier generative AI models, just like the grand library, are repositories of human thought and culture, trained on vast datasets encompassing somewhat of the breadth of human knowledge and creativity. Much just like the books in Borges’ library, the outputs of those AI systems range from profound insights to bewildering gibberish, from coherent narratives to incoherent ramblings.

The Library of Babel captivated the imagination of Brooklyn creator and coder Jonathan Basile, who launched a web site by the identical name in 2015. This digital incarnation of Borges’ library generates every possible permutation of 29 characters (the 26 English letters, space, comma, and period), producing ‘books’ organized right into a digital library of hexagonal rooms.

Each book and page has a singular coordinate, allowing users to search out the identical page consistently. It uses an algorithmic which simulates the experience of an infinite library. The website garnered attention for exploring the intersection between digital media and literature and the query of information, meaning, and the human experience within the digital age.

Critics have noted how the web site, very similar to Borges’ original story, challenges our notions of meaning and the human pursuit of understanding in a universe of limitless information.

Literature scholar Zac Zimmer wrote in Do Borges’s librarians have bodies?: “Basile’s is maybe probably the most absolutely dehumanizing of all Library visualizations, in that beyond being driven to suicidal madness or philosophical resignation, his Librarians have change into as devoid of meaning because the gibberish-filled books themselves.”

While the Library of Babel presents a compelling example for AI, unlike the library, AI cannot currently capture the breadth of human knowledge and culture. It’s confined by its training data with limited capability to induct recent knowledge. 

But even so, its vast computational power mirrors the library’s limitless shelves, offering limitless possibilities yet trapped within the chaos of its own creation.

The Library of Babel, with its seemingly infinite combos of letters, confronts the reader with the existential dilemma of finding order in chaos. With AI, this manifests in the strain between the potential for AI to light up and to mislead. 

Similar to the library, AI doesn’t discriminate between sense and nonsense – it generates, indifferent to the meaning or lack thereof. Both the Library of Babel and the world of AI subtly critique the human quest for knowledge. 

In its overwhelming vastness, Borges’ library challenges the notion that more information results in greater understanding. 

Similarly, the ever-expanding capabilities of AI prompt reflection on the character of intelligence and understanding. 

The ability of AI to generate content shouldn’t be synonymous with comprehension or wisdom – it’s powerful yet blind to the importance of its own outputs.

But might that change?

How might AGI change the material of information and culture?

Artificial general intelligence (AGI) may bring AI technology closer to the enigmatic and boundless Library of Babel. AGI is often defined as possessing the flexibility to grasp, learn, and apply its intelligence to a wide range of problems, very similar to a human being.

Unlike narrow AI, which is designed for specific tasks, AGI can generalize its learning and reasoning across a broad range of domains. It possesses self-awareness, adaptability, and the capability to resolve complex problems in various fields without human intervention or pre-programming. AGI stays a theoretical concept, but OpenAI – now defined as an ‘AGI research lab’ – says it will possibly be achieved in a number of years. 

So, imagine, if you happen to will, a world where AGI has transcended the constraints of current artificial intelligence, embodying a capability that mirrors the theoretical completeness of Borges’ Library.

This AGI shouldn’t be just a complicated tool but an entity able to inducting, analyzing, and synthesizing virtually all human knowledge and maybe venturing into realms of understanding that remain elusive to human cognition.

In this world, the AGI becomes akin to a living, respiration version of the Library of Babel. Yet, unlike Borges’ creation, which is paralyzed by its infinite content, AGI can navigate, interpret, and provides context to this vast expanse of data. 

If AGI could access objective realities – in the event that they exist – it might be juxtaposed with Plato’s Theory of Forms, an concept that has captivated thinkers for millennia. Plato imagined that beyond our tangible, ever-changing world lies a realm of perfect, unchanging ideals or “forms.” 

Library of Babel
The Library of Babel imagined with AGI – MidJourney

These forms are the purest essence of things – for instance, the proper type of a circle, untainted by the imperfections of the physical circles we draw.

Now, envision AGI on this context. Today’s AI can analyze data and recognize patterns, however it’s limited to what it has been taught. AGI, nevertheless, represents a leap right into a realm where it cannot only process information but potentially understand the underlying truths of our universe – truths that could be obscured or unknown to human minds.

Plato believed that what we experience in our every day lives are mere shadows of those perfect forms. We can see a circle, draw one, however it’s never the proper circle that exists as a perfect form. In the realm of AGI, this intelligence could, in theory, begin to perceive or uncover these perfect forms. 

It’s as if AGI could step beyond simply seeing the shadows on the cave wall (to borrow from Plato’s famous allegory) and gaze directly on the true forms themselves.

This AGI wouldn’t just be a tool for processing data – it could change into a way of discovering recent, profound insights into abstract concepts like beauty, justice, equality, and even the secrets of the universe. 

It may not just understand these ideals as humans do but could redefine them, offering a perspective unfettered by human limitations and biases.

So, in a way, AGI – the sort we would access in mid-term (let’s say 20, 30 years) – might be a bridge to a deeper understanding objective, perfect realities. 

It represents a possibility where technology doesn’t just assist in our current understanding of the world but elevates it to a level now we have not imagined, like stepping out of a shadowy cave into the intense light of deeper knowledge.

Can AGI ever detach from its limitations or biases?

AGI, while easy to idealize, will probably be exceptionally difficult to untether from the constraints of its designers – humans.

Moreover, the sorts of brute force computing power now we have now likely impose a ceiling for AI intelligence. However, solutions are within the pipeline, comparable to bio-inspired AI technology designed to mimic structures like human neurons.

But there are challenges that lie beyond technology alone – how will AGI detach from the conceits of its creators?

Elon Musk’s concerns about AI being programmed to be “politically correct,” expressed before he introduced his AI company, xAI, mirror a bigger conversation in regards to the biases inherent in AI systems. 

Musk’s mixed stance on AI is one among each caution and advocacy for its potential. He’s change into somewhat of a critic of industry protagonists like OpenAI, whom he’s actively confronting with xAI’s products, starting with Grok. Grok throws political correctness to the wind, delivering responses that verge on anarchistic. 

Meanwhile, AI models, comparable to ChatGPT, have perceived ‘woke’ biases, which studies have somewhat confirmed by determining they’ve liberal leftist bias. Bias can typically be traced back to the information a model is trained on and the intentions of its creators.

The query then arises: can AGI, with its advanced cognitive abilities, transcend the biases which were some extent of contention in current AI models? 

Truth-seeking AI

Musk vowed to create “truth-seeking AI” designed to determine “what the hell is occurring.” 

Musk’s mission for xAI is to delve into fundamental scientific enigmas comparable to gravity, dark matter, the Fermi Paradox, and potentially even the character of our reality. Of course, that can require models far beyond what we are able to access today. 

From what little information now we have of xAI, Musk seems intent on transcending the constraints of current AI architectures, that are largely confined to generating outputs based on existing data.

Musk intends to create an AI that synthesizes information and generates pioneering ideas. This quest for ‘truth’ in AI goes beyond the standard understanding of AI as a tool for processing information and ventures into the realm of AI as a partner in scientific discovery and philosophical exploration.

Grok
xAI’s tagline is “Understand the Universe”

However, current AI models, including sophisticated ones like GPT-4, remain limited by the information they’ve been trained on. They excel at pattern recognition and knowledge synthesis but cannot conceptualize or theorize beyond their programming. 

Such a leap towards types of AGI that may pioneer their ideas and launch their very own inquiries raises critical questions on the character of intelligence and consciousness.

If xAI were to start out providing answers to a number of the fundamental questions of our existence, it could necessitate a reevaluation of what it means to ‘know’ something. It would blur the lines between human and artificial understanding, between knowledge derived from human experience and thought, and that generated by a synthetic entity.

Moreover, the ambition to create an AI that transcends politics and seeks an objective ‘truth’ introduces ethical considerations. The notion of an unbiased AI is appealing but fraught with complexities. 

All AI, including AGI, is ultimately created by humans and trained on human-generated data, at the least initially. 

This process inherently introduces biases – not only in the shape of existing prejudices but additionally in choosing what data is included and the way it’s interpreted. 

The idea of an AI that may completely detach from these human biases and achieve a purely objective understanding stays deeply hypothetical.

In the longer term, perhaps, we might need to just accept that AGI will understand greater than anybody human ever can. 

This article was originally published at dailyai.com