The AI age has begun in the year of the dragon


Advertisement: Click here to learn how to Generate Art From Text

If you were hoping that the world would get over AI fever in 2024, you are going to be sadly mistaken. The advancements in hardware and software (everywhere), are opening the floodgates for dynamic applications. AI is a powerful tool that can be used to generate new ideas. that suggest that 2023 was the year where we only really began to scratch the surface.  

This year — the Year of the Dragon in the Chinese Zodiac — will see a widespread and strategic  integration of Gen AI in all sectors. With risks assessed and strategies beginning to take shape, businesses are poised to leverage gen AI not just as a novel technology, but as a core component of their operational and strategic frameworks. In short, CEOs and business leaders, having recognized the potential and necessity of gen AI, are now actively seeking to embed these technologies into their processes.  

The resulting landscape is one where gen AI becomes not just an option, but an essential driver of  innovation, efficiency and competitive edge. This transformative shift signifies a move from tentative exploration to confident, informed application, marking 2024 as the year where gen AI transitions from an Emerging TrendThis is a basic business practice. 

Volume and variety

A key dimension is the growing understanding of how gen AI allows for both increased volume and variety of applications, ideas and content.  

VB Event

The AI Impact Tour – NYC

We’ll be in New York on February 29 in partnership with Microsoft to discuss how to balance risks and rewards of AI applications. Please fill out the form below to request an invitation to this exclusive event.


Request an invitation

The staggering amount AI-generated content will have ramifications that we are only beginning to  discover. Due to the sheer volume of this content (since 2022, AI users have collectively created more than 15 billion images — a number which previously took humans 150 years to produce), historians will have to view the internet post-2023 as something completely different to what came before, similar to how the atom bomb set back radioactive carbon dating.  

However, regardless of what gen AI is doing to the internet, for enterprises, this expansion is elevating the standard for all players across all fields, and signals a critical juncture where not engaging with the technology may not just be a missed opportunity, but a competitive disadvantage. 

The jagged border

In 2023, we learned that gen ai not only raises the bar across industries, but in employee capabilities. In a study by YouGov in the last year90% of workers said they were satisfied with their jobs. AI is improving their productivity. One in four of respondents use AI on a daily basis (with 73% of  workers using AI at least once a week).  

Separate studies found that with the right training, employees completed 12% of tasks 25% faster with the help of gen AI, and that overall work quality rose 40% — with those of lower skill level making the most gains. However, for tasks outside AI’s  capabilities, employees were 19% less likely to produce correct solutions.  

This duality has led experts to call it the “jagged frontier”AI capabilities. This works as follows: On one end of the spectrum, we witness AI’s remarkable prowess — tasks that once seemed insurmountable for machines are now executed with precision and  efficiency. 

Yet, on the flip side, there are tasks where AI falters, struggling to match human intuition and  adaptability. These are areas marked by nuance, context and intricate decision-making — realms  where the binary logic of machines (currently) meets its match.

Cheaper AI

This year, as enterprises begin to grapple and master the jagged frontier, we will see gen AI projects start to land and become normalized. Underlying this adoption is the decline in the cost of training foundational large language models (LLMs)The silicon optimization has improved (by about half every two months). 

Together with increased demand and amidst global shortages, the AI chip market is looking to become more affordable in 2024, as alternatives to industry-leaders like Nvidia emerge from the woodwork.  

Likewise, new fine tuning methods that can grow strong LLMs out of weak ones without the need for additional human-annotated data — such as Self-Play fIne-tuNing (SPIN) — are leveraging synthetic data to do more with less human input. 

Enter the ‘modelverse’

This reduction in cost is opening doors for a wider array of companies to develop and implement  their own LLMs. The implications are vast and varied, but the clear trajectory is that there will be a surge in innovative LLM-based applications over the next few years.  

In 2024 we will also begin to see a move away from AI models that are primarily cloud-based to AI that is executed locally. This evolution is driven partly by hardware advancements like Apple Silicon, but it also capitalizes on the untapped potentials of raw CPU power in everyday mobile devices. 

In business, the same is true. Small language models (SLMs). are set to become more popular across large and medium-scale enterprises as they fulfill more specific, niche needs. As their name suggests, SLMs are lighter in weight to LLMs — making them ideal for real-time applications and  integration into various platforms.

So, while LLMs are trained on vast amounts of diverse data, SLMs are trained on more domain-specific data — often sourced from within the enterprise —  making them tailored to specific industries or use cases, all while guaranteeing relevance and  privacy.  

Large Vision Models (LVMs), a shift in the use of large vision models

As we transition into 2024, the spotlight will also shift from LLMs towards large vision models (LVMs) — particularly domain-specific ones — that are set to revolutionize the processing of visual data. 

While LLMs trained on internet text adapt well to proprietary documents, LVMs face a unique challenge: Internet images predominantly feature memes, cats and selfies, which differ significantly from the specialized images used in sectors like manufacturing or life sciences. Therefore, a generic LVM trained on internet images may not efficiently identify salient features in specialized domains. 

However, LVMs tailored to specific image domains, such as semiconductor manufacturing or pathology, show markedly better results. Research demonstrates that adapting an LVM to a specific domain using around 100K unlabeled images can significantly reduce the need for labeled data, enhancing performance levels. These models, unlike generic LVMs, are tailored to specific business domains, excelling in computer vision tasks like defect detection or object  location. 

Businesses will start to adopt the technology elsewhere Large graphical models. These models excel in  handling tabular data, typically found in spreadsheets or databases. They stand out in their ability  to analyze time-series data, offering fresh perspectives in understanding sequential data often found in business contexts. This capability is crucial because the vast majority of enterprise data falls into these categories — a challenge that existing AI models, including LLMs, have yet to  adequately address. 

Ethical dilemmas

These developments will need to be supported by a rigorous ethical framework. Common consensus is that we got previous general purpose technologies (technologies that have broad-based applications, profoundly impact diverse areas of human activity and fundamentally change the economy and society) very wrong. While presenting immense benefits, tools such as the smartphone and social media also came with negative externalities that permeated all facets of our lives, whether or not we engaged with them directly. 

With gen AIRegulating is considered essential to avoid repeating past mistakes. However, it may fail, stifle innovation or take time to go into effect, so we will see organizations opposed to governments leading the regulatory charge. 

Perhaps the most well known ethical quagmire gen AI introduced last year was the issue of copyright. As AI technologies advanced rapidly, they brought to the fore pressing questions about intellectual property rights. The crux of the issue, of course, lies in whether and how AI-generated content, which often draws upon existing human-created works for training, should be subject to copyright laws. 

The AI/copyright tension exists because copyright law was created to prevent people using other  people’s IP unlawfully. Reading articles or texts for inspiration is allowed, but copying it is not. If a person reads all of Shakespeare and produces their own version, this is considered inspiration, yet the challenge is that AI can consume limitless volumes of data, as opposed to a human-constricted limit.  

The copyright/copywrong controversy is only one facet of the media in flux. In 2024, we will see the result of landmark, precedent-setting cases such as the OpenAI vs. NYT (however, it is unclear if this will ever go to trial or is simply a bargaining tool by the publisher) and witness the ways in which the media landscape adapts to its new AI reality. 

Deepfakery is on the rise

In terms of geopolitics, the AI story of the year will inevitably be how this technology is intersecting with the biggest election year in human history. This year, more than half of the world’s population are heading to the polls, with presidential, parliamentary and referential votes scheduled in nations including the U.S., Taiwan, India, Pakistan, South Africa and South Sudan. 

Bangladesh, where the January elections were held, has already experienced this kind of interference. Influencers and media outlets that support the government actively promoted disinformation using AI tools. 

In one instance, a deepfake video (that was subsequently taken down) showed an opposition figure appearing to retract support for the people of Gaza, a stance that could be detrimental in a nation where the majority of Muslims hold a strong solidarity with Palestinians. 

The threat posed by AI imagery is real. Recent research revealed that subtle changes designed to deceive AI in image recognition can also influence human perception. The findings, published in Nature Communications, underscores the parallels between human and machine vision — but more importantly, it highlights the need for more research into the impact of adversarial images on both people and AI systems. These experiments showed that even minimal perturbations, imperceptible to the human eye, can bias human judgments, akin to the decisions made by AI models. 

While a global consensus is emerging around the concept of watermarking (or content credentials) as a means to distinguish authentic content from synthetic, the solution is still fraught with its own complexities: Will detection be universal? If so, how can we prevent people from abusing it — labeling work that is synthetic when it is not? On the other hand denying everyone the ability to detect such media cedes significant power to those with it. Once again, we will find ourselves asking: Who gets to  decide what is real?

With public trust across the world remaining firmly at a nadir, 2024 will be the year when the world’s biggest election year intersects with the most defining technology of our time. For good and for bad, 2024 marks the year wherein AI is applied in real, tangible ways. Hold on tight.

Elliot Leavy founded ACQUAINTED, Europe’s first generative AI consultancy.


Welcome to the VentureBeat Community!

DataDecisionMakers allows experts, including those who work with data, to share their data-related insights and innovations.

Join us at DataDecisionMakers to learn about the latest ideas, information, and best practices in data and data technology.

You might even consider Contributing an article of your own!

Read more from DataDecisionMakers

‘ Credit:
Original content by “Year Of The Dragon: We Have Entered the AI Age”

Read the full article here ‘

Leave a Reply

Your email address will not be published. Required fields are marked *