Tag Archives: Generative AI

[Google I/O 2023] Google Bard – What is That?

After Google I/O 2023 last week, you might have noticed that your Android smartphone pushing a notification to you. It is a prompt for you to try Google’s updated Bard. Most of you on Google’s email platform (Gmail) might also get an email asking you to try Bard today. If you are familiar with AI (artificial intelligence) news, you might already be familiar with Google’s Bard alongside OpenAI’s ChatGPT. To those, it might sound like a foreign object.

In simple terms, Google Bard is really the Google version of ChatGPT. While ChatGPT is developed by OpenAI, Bard is completely Google. You want to keep in mind that both ChatGPT and Bard are two separate platforms altogether though before jumping to conclusions and say that they are the same things. They are both categorised as generative AI, but they are both different from one another.

Unlike ChatGPT which has existed for some time, and is in its fourth iteration, Google Bard is fresh out of the oven; two months out of the oven, to be fair. Like ChatGPT, Google Bard was launched as an experiment. Like ChatGPT as well, the technology for Google Bard is not exactly new.

What is Google Bard?

Screenshot 2023 05 15 162043
Source: Google

As mentioned, Google Bard is a generative and creative AI by Google. Instead of overcomplicating the explanation, Google’s FAQ says that Google Bard is technically based on their LaMDA (Language Model for Dialogue Applications) AI model, Google’s very own linguistics program written for conversational purposes. When we say conversational, we do not mean that it will be like a regular conversation with a human being, but LaMDA aims to make it close.

To be fair, Google’s conversational AI is not something you have not seen before, you see it with Google Assistant whenever you call out “Hey, Google,” or “Okay, Google”. You can even use Google’s clever Assistant to get you a booking for a restaurant by having Google Assistant make the call and get the booking done, instead of you calling the restaurant yourself. In their demo a few years ago, Google’s Voice Assistant sounded so natural that the other person on the other end of the line could not even tell that they are speaking to an artificial person. This proves that LaMDA works, and has a place in the world. But our many use case of the Google Assistant even with Google Nest systems is prove enough that conversational AI has many uses in the current world.

Bard is not just a conversationalist though. It is more than that, a generative AI of sorts. It still has its roots in LaMDA, but it is a lot more than that now. It is made as a collaborative tool, for you to basically generate ideas, tabulate and make sense of data, help you plan things, help you design tools and steps, collate your calendars, and even use it as a learning tool.

According to Google, Bard is made to create original contents at the request and behest of individual users. Meaning that the algorithm could be different are results can be different from one person to another. Because it is Google, any request or question you post to Bard might prompt Bard to look into hundred or thousands of sources and draw conclusions, or present result in a way the does not infringe copyright or plagiarism laws. In the case that it does take up contents from another source, Bard will acknowledge and cite its sources. Google Bard is not built to write your college essay though, it is built to be a collaborator to manage your work and your life, to make it more seamless somehow over just Googling things. They do actually have a ‘Google It’ button for you to make full use of Google’s search engine though.

It is not a 100% solution for your own research and use case though. Google has mentioned and stressed that Google Bard is an experiment. It is an opportunity for their AI engines to learn even more at an accelerated pace with public input and use. Google Bard is meant to be iterated, which also means that the current form of Google Bard will not be final. They also mention that Google Bard, at its current form will not be 100% accurate at all times; hence, the ‘Google It’ button on Bard. While it is open source, Google also says that Bard is not meant to be used commercially or for advertising purposes at this time.

Why Bard?

Screenshot 2023 05 15 162312
Source: Google

The entire existence of Bard could be a sharp response to OpenAI’s ChatGPT. The success of the open-source AI platform has sort of forced Google to quickly introduce their own AI tool for use to the public. If they are to be believed, Google could offer the most powerful AI tool for the masses.

In the recent Google I/O 2023, Google has officially embraced Bard and announced that they have moved Bard to PaLM 2, an improved language model that offers more capabilities of Google Bard compared to just conversational based on LaMDA model. PaLM 2 now offers Bard the ability to code and program. It also allows Bard to solve even more complex mathematical problems and process through more complex reasoning models that offers Bard the ability to make better decisions over time.

As of Google I/O 2023, Google has opened the Bard experiment to more than 180 countries as of writing and is available in Japanese and Korean. As things go, Google is planning to open the experiment to more regions and make Bard available in about 40 languages. On top of more languages and regions, where the older Google Bard was mostly just conversational via text, the new improvement at Google I/O 2023 adds some visual flavours to your conversations with Bard. They have integrated Goole Lens into Bard and allow you to now scan photos of your things at home and let Bard come up with whatever captions you might want. You can even add photo references to your Google Bard generated itinerary when you travel.

But it is not just the surface updates for Google Bard. For Google I/O 2023, they have announced that Bard is not just a tool that is isolated from any other systems. Google is making the Bard available with an “export” button for collaboration purposes in the form of exporting and running codes on Python. You could directly copy email responses into your Gmail or Google Docs, if you want. If you want more out of Bard, you can even expect Adobe Firefly integration in the coming future for even more powerful generative tools like complete poster designs based on both Google’s and Adobe’s combined algorithms. They have also announced that they are working with more partners like Kayak, OpenTable, ZipRecruiter, Instacart, Wolfram and Khan Academy to get their Google Bard project integrated into their services and products.

In this case, where OpenAI is allowing you to plug its API anywhere and get it working with minor tweaks, Google is not looking to just do that. Google is offering deep integration with their partners to create even more, to become an even more powerful tool in your toolkit for the future. They look to open up even more opportunities and applications for the average user with deeper and more curated collaborations with partnering brands. While that may not necessarily be the best thing to do for some, it is a way forward for more integrated services and solutions to serve individuals and businesses better. It even allows partnering companies to understand their users and customers better in some cases.

Adobe Firely, the Next-Generation AI Made for Creative Use

AI (Artificial Intelligence) generated graphics is not a new thing. You have things like OpenArt and Hotpot these days where you can just type in the keywords to the image you want, and let the engine generate art for your use. Even before AI generated graphics though, the implementation of AI within the creative industry is nothing new. NVIDIA has used their own AI engine to write an entire symphony, and even to create 3D environments using their Ray Tracing engines. Adobe too have something they call the Sensei. They the AI tool is implemented across their creative suite to understand and recognise objects better, fill details where needed more naturally, and even edit videos, images, or even texts quickly and efficiently. Now, they have Firefly.

Firefly is not a new separate AI system from Adobe’s Sensei. Firefly is a part of a larger Adobe Sensei generative AI together with technologies like Neural Filters, Content Aware Fill, Attribution AI, and Liquid mode implemented across several Adobe platforms. Unlike those platform specific implementations though, Adobe is looking to put Firefly to work on a number of various platforms across their Creative Cloud, Document Cloud, Experience Cloud, and even their Adobe Express platforms.

So, what is Adobe Firefly? We hear you ask. It is technically Adobe’s take on what a creative generative AI should be. They are not limiting Firefly to just image generation, modification, and correction. It is designed to allow any sort of content creators create even more without needing to spend hundreds of hours to learn a new skill. All they need to do is to adapt Firefly in their workflow and they will get contents that they have never been able to create before, be it images, audio, vectors, texts, videos, and even 3D materials. You can have different contents every time too with Adobe Firefly; the possibilities, according to Adobe, are endless.

What makes Adobe’s Firefly so powerful is the power of the entirety of Adobe’s experience and database behind it. Obviously Adobe’s Stock images and assets is a huge enough library for the AI implementation to dive into. The implementation can also look into using openly licensed assets and public domain contents in generating its contents. The tool, in this case, will prevent any IP infringements and help you avoid plenty of future litigations.

Adobe Firefly Cover
Source: Adobe

As Firefly is launched in its beta state, it will only be available as an image and text generator tool for Adobe Express, Adobe Experience Manager, Adobe Photoshop, and Adobe Illustrator. Adobe plans to bring Firefly into the rest of their platforms where relevant in the coming future. They are also pushing for more open standards in asset verification which will eventually include proper categorization and tagging of AI generated contents. Adobe is also planning to make the Firefly ecosystem a more open one with APIs for its users and customers to integrate the tool with their existing workflows. For more information on Adobe’s latest generative AI, you can visit their website.

Be A Maestro with AWS DeepComposer

You would think that when it comes to making compositions and music, you’d need a really good ear and knowledge of the arts. Not so much with Amazon Web Service’s new AI (Artificial Intelligence) service focused on creating musical pieces with a keyboard! DeepComposer is the latest in a series of Machine Learning focused services that AWS has introduced since it’s announcement of DeepLens at Re:Invent 2017.

The new music based AI is a 32 key, 2 octave keyboard which will allow developers to familiarise themselves with using Generative AI. The simple application of Generative AI in DeepComposer will take short riffs and generate a full compositions.

A brief diagram explaining how AWS’s DeepComposer works. (Source: AWS)

The DeepComposer generative AI will be able to layer and generate songs based on pre-trained models or even user defined models. The pre-trained models are able to generate based on algorithms developed by training the AI with large musical data sets. The user defined models give users better control of the generative AI. Users will be able to define multiple parameters including the Architecture and Discriminator. The latter allows the AI to distinguish between the genres and determine the overall composition.

Announcing AWS DeepComposer with Dr. Matt Wood, feat. Jonathan Coulton

Being a machine learning model, DeepComposer is continually learning to identify music types. The AI will improve with time as it learns and generates more music based on the models and riffs. It will also be able to generate music which mimics a defined model. Amazon’s release touts, ” you have to train as a counterfeiting expert in order to become a great counterfeiter “.

DeepComposer isn’t just linked to the physical keyboard. It also has a digital keyboard interface which allows users to compose on the go. Using this approach, AWS is hoping that Generative AI models are made more approachable for those looking to explore their applications.

The new feature is currently available for preview on AWS at the DeepComposer website. Also on the website is a FAQ to address some of the questions that new users may have.